var/home/core/zuul-output/0000755000175000017500000000000015157452621014535 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015157464300015476 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000214201415157464115020263 0ustar corecoreMhikubelet.log_o[;r)Br'o -n(!9t%Cs7}g/غIs$r.k9Gf ?lEZ͖o_˖_" ^^^oo}͛ji/hxK}l11O6EYn*jn獼خx~̖K^_/_p/Jz\,W]EoO/(̗?<x Ζbx= x% +#4^ 8D^ώI8&xėf9E៾|3F}Zl⇓8T*v (6pk**+ Le*gUWi [ӊc*XCF*A`v cXk?`QlrTvb)EZW3)7ɀ;$#LcdHM|J^[^Sg\W#NqɌDSd1d9nT#Abn q1J# !8,$RNI? r.|I"n j/\U R[EC 7g/7_E'c/z&BBb}1lχtO Ң`?Tӣ 5W=Xz`̷~F<5a|X&pi]>yE*,?k 9Z29}}(4ҲIFyG -^76ox7,*uvf d |TRZ;j?| |!I狓 3w̗`{K0Aȶ9W E%*mG:toG(;h0!}j)CMitmy߀~s{@Q,}s=LN YlYd'Z;.K'~_.Vp|A*Z*}QJ0SqAYE0i5P-$̿:R€ N0RQGkuWZ^yhi-cS4 6"mKaFרfq&}͕C,RizpV:!җӣ D>P.BvJ>mIyVVTF% tFL-*$tZm2AČAE9ϯ~ihFf&6"W&\jVJ5&jNgB|90v߁R:~U jڞU~oN9菏xԞJ|d`V)F5d,0SSNK9ް4:ÒozsB<^/鄌4:B%cXhK I}!5 YM%<>"۞)Za@Ι}YJz{ɛrm4^jC d-saܺCY "D^&M){ߙ>:i V4nQi1h$Zb)ŠȃAݢCr|<~gQwQ!q/C>*({bʂ!0[r_G{j 6JYǹ>zs;tc.mctie:x&"bR4S uV8?,0-Y8Uav0NET݃jYAT` &AD]Ax95mvXYs"(A+/+o+{b]}@UP*5ì"Ml؈W|sN{mL=@N'DǭZrb5Iffe6Rh&C4F;D3T\[ bk5̕@UFB1/ {f'}KXg%q3Ifq CXReQP2$TbgK ؾ#AZ9 K>UHkZ;oﴍ8M*a~ff~6|Y,d,`!qIv꜒"T[1!I!NwLv}\|s.|=o4_P\(Lۻ5_vJBR_v'5n[lGhNU˿ۜ6C9OC7Osnkje*;iΓF[^>؃n糨I@[ tWv Few9J֥WmN^<.eܢMρ'JÖŢո%gQ=p2YaI"&ư%# y}ùXz!bm5uAߙXc90ov,{*9ߎ% qƦat:F=uNvD=mdZB4']a.QO:#'6RE'E3 */HAYk|z|ؾPQgOJӚ:ƞŵ׉5'Hdg;md^6%rd9#_~2:Y`&US tDkQ;>" ؾ:9_))wF|dߗXiTcLQMhg:F[bTm!V`AqPaPheUJ& z?NwGj{VjQS,؃I'[9;|y]?z碨T u]68 QeC Hl @R SFZuU&uRz[2(A1ZK(O5dcXxP*]"U!2`i*6J'2I+6֗%)FV[/GWM\jP' HT؋ BvaijH{=]b~3Z|]ɍ%ʺsl!C>Egl1$9  ֲQ$'dJ<6=bs^)'Hz ȍq!a}P `x ăil_=!X;__ZNۈO;j -ݙrbH9t֥s`Mh,욕v`;VI&ϊх{211 VSxSew{~n}Lr7+Tcj.R2>e3V#b,PY0TEu1L/]LTB4$`H6NI\b*AYA\(u|@ [h-,j7gDTÎ4oZ$j!.;嶑, }t&&\5u17\I@ 5O? ʴ(aPqPϟ'K0D"\KjPQ>Y{Ÿ>14`SČ.HPdp12 (7 _:+$ߗv{wzM$VbήdsOw<}#b[E7imH'Y`;5{$ь'gISzp; AQvDIyHc<槔w w?38v?Lsb s "NDr3\{J KP/ߢ/emPW֦?>Y5p&nr0:9%Ws$Wc0FS=>Qp:!DE5^9-0 R2ڲ]ew۵jI\'iħ1 {\FPG"$$ {+!˨?EP' =@~edF \r!٤ã_e=P1W3c +A)9V ]rVmeK\4? 8'*MTox6[qn2XwK\^-ޖA2U]E_Dm5^"d*MQǜq؈f+C/tfRxeKboc5Iv{K TV}uuyk s" &ﱏҞO/ont~]5\ʅSHwӍq6Ung'!! e#@\YV,4&`-6 E=߶EYE=P'ސlnbu;Bx`C:gev\g7 6܋ hH+P5co.Q/cC@.sABC{تI=;̶n2¦l::s@ -Mn3䦴mHЭj !'T9Xsl o:d lzzMvYź ^ ٲAPm쪊m\9htwmjQ\c5&,|^C.SS P󂏛o n8Fkb^s/&a[s~W &ɿ^\r\ߺnqZV@z%=\#|-3ڝa$ΫM|-LsXY r# v&讳YE 6X̀v"@L'aEN^8 n`т ti6{b?-X;|iDɈP͐b7jk *bmc`  SgkmOl7^~xAE,Pmqs;l};Щ۸l?28Ćn.I0Yhu ;ZeY}Qg?lvחzäTC 4zv)|Vy7߯@qC cN ͯ~1-b }kAn=)m 3fo˶_ XJNC5B~6%d+, Z`ͲH-nမ^WbPFtOfD]c9\w+ea~~{;Vm >|WAޭi`HbIãE{%&4]Iw Wjoru ݜmKnZ<X; ۢ( nx K8.|DXb +*598;w)zp:̊~;͞)6vnM!N5Cu!8Wq/`FUwWAֻ,Qu W@ Fi:K [Av*_958]a:pmQ&'ᚡmi@ zF(n&P;)_]µ!doR0`pl`~9Fk[ٺ+4Hhao-jϸ??R<lb#P-^39T|L /~p│x@Bq"M/lja\b݋af LnU*P(8W[U6WX ZoѶ^SH:K:%Qvl\b FqQI.ȨHWo;Nw$͹O$oEE-eq=.*Dp,V;(bgJ!gF)892sw*+{[or@x,))[o新#.͞.;=fc<)((b۲Eumw峛M2,V[cm,S~ AF~.2v?JNt=O7^r.@DEuU1}g$>8ac#sĢB\PIPfwJQJ;Qxm &GBf\ZA$Ba-z|A-I @x70 晪MV)m8[6-Te@`E|=U D(C{oVa*H7MQK"<O%MTTtx袥:2JޚݶKd7UZihRk71VDqiގ\<:Ѓ3"gJJčE&>&EI|I˿j2ǯɘCGOa9C1L ={fm&'^tigk$DA' elW@Tiv{ !]oBLKJO*t*\n-iȚ4`{x_z;j3Xh ׄ?xt.o:`x^d~0u$ v48 0_ | E"Hd"H`A0&dY3 ً[fctWF_hdxMUY.b=eaI3Z=᢬-'~DWc;j FRrI5%N/K;Dk rCbm7чsSW_8g{RY.~XfEߪg:smBi1 YBX4),[c^54Sg(s$sN' 88`wC3TE+A\.ԍל9 y{͝BxG&JS meT;{З>'[LR"w F05N<&AJ3DA0ʄ4(zTUWDdE3̻l^-Xw3Fɀ{B-~.h+U8 i1b8wؖ#~zQ`/L 9#Pu/<4A L<KL U(Ee'sCcq !Ȥ4΍ +aM(VldX ][T !Ȱ|HN~6y,⒊)$e{)SR#kהyϛ7^i58f4PmB8 Y{qeφvk73:1@ƛ.{f8IGv*1藺yx27M=>+VnG;\<x7v21՚H :[Γd!E'a4n?k[A׈(sob 41Y9(^SE@7`KIK`kx& V`X0,%pe_ן >hd xе"Q4SUwy x<'o_~#6$g!D$c=5ۄX[ു RzG:柺[ӏ[3frl ô ހ^2TӘUAT!94[[m۾\T)W> lv+ H\FpG)ۏjk_c51̃^cn ba-X/#=Im41NLu\9ETp^poAOOgJ8@o2k'Hr~4Z(I8!H G8HNW%1Tќ^?xBVQXodՔz q[*ڔC"1Ȋ-R0ڱ}oF4 3vFf#8^Vє+k@ :)@%9@nA B q 62!/ 6G (" u:)fSGAV(e֖t܁ ft~c.!R0N<R{mtdFdHÃФsxBl] " Δ<=9i/ d ␙F9Ґ)Hnxps2wApP!se]I)^ k?'k:%Ѹ)?wɧ6a{r7%]_Ϧi~ԞnZhubW*IakVC-(>Z#"U4Xk1G;7#m eji'ĒGIqB//(O &1I;svHd=mJW~ړUCOīpAiB^MP=MQ`=JB!"]b6Ƞi]ItЀ'Vf:yo=K˞r:( n72-˒#K9T\aVܩO "^OF1%e"xm뻱~0GBeFO0ޑ]w(zM6j\v00ׅYɓHڦd%NzT@gID!EL2$%Ӧ{(gL pWkn\SDKIIKWi^9)N?[tLjV}}O͌:&c!JC{J` nKlȉW$)YLE%I:/8)*H|]}\E$V*#(G;3U-;q7KǰfξC?ke`~UK mtIC8^P߼fub8P銗KDi'U6K×5 .]H<$ ^D'!" b1D8,?tT q lKxDȜOY2S3ҁ%mo(YT\3}sѦoY=-- /IDd6Gs =[F۴'c,QAIٰ9JXOz);B= @%AIt0v[Ƿ&FJE͙A~IQ%iShnMІt.޿>qƽl' ,Τ9)%@ wl42iG.y3bBA{pR A ?IEY ?|-nz#}~f ‰dŷ=ɀ,m7VyIwGHέ 2tޞߛM{FL\#a s.3\}*=#uL#]  GE|FKi3&,ۓxmF͉lG$mN$!;ߑl5O$}D~5| 01 S?tq6cl]M[I5'ոfiҞ:Z YՑ"jyKWk^dd@U_a4/vvV qHMI{+']1m]<$*YP7g# s!8!ߐ>'4k7/KwΦθW'?~>x0_>9Hhs%y{#iUI[Gzďx7OnuKRv'm;/~n-KI`5-'YݦD-!+Y򼤙&m^YAKC˴vҢ]+X`iDf?U7_nMBLϸY&0Ro6Qžl+nݷ" 㬙g|ӱFB@qNx^eCSW3\ZSA !c/!b"'9k I S2=bgj쯏W?=`}H0--VV#YmKW^[?R$+ +cU )?wW@!j-gw2ŝl1!iaI%~`{Tռl>~,?5D K\gd(ZH8@x~5w.4\h(`dc)}1Kqi4~'p!;_V>&M!s}FDͳ֧0O*Vr/tdQu!4YhdqT nXeb|Ivż7>! &ĊL:}3*8&6f5 %>~R݄}WgѨ@OĹCtWai4AY!XH _pw騋[b[%/d>. !Df~;)(Oy )r#.<]]i-*ػ-f24qlT1  jL>1qY|\䛧\|r>Ch}Ϊ=jnk?p ^C8"M#Eޑ-5@f,|Ά(Շ*(XCK*"pXR[كrq IH!6=Ocnи%G"|ڔ^kПy׏<:n:!d#[7>^.hd/}ӾP'k2MؤYy/{!ca /^wT j˚ب|MLE7Ee/I lu//j8MoGqdDt^_Y\-8!ד|$@D.ݮl`p48io^.š{_f>O)J=iwwӑ؇n-i3,1׿5'odۆ3(h>1UW蚍R$Wv8FHӜ"D$aǽO8'1lfYuB!aMHVuV_K2k*`cKxuBG&24T}Lai 0Va(7K#ӊ!,ZDxFQO*lם>!4ӥ2 ]8â6 U`V%`!c%؎ʨTzrKh! c.}.D>)d_ 8rcu,wf2?Ǡ*_lDn}rauyFp*ɨ:UiM2r:9ct X1lmĪ o玓,R%!`hGT LYF#g<cm${|Xdu4tmtїUJ\~dc0KcMlf2?mμQ ߉J4WrSHTdp"ӹ'cJq2zPlX̯.0H!ND@UapVoGڧD5>H]f@!=߸2V%Z 0"G4ȇʩ@]>Y$ًF_Mm_Tt)ib+q&EXFu򾬳ǝ/RS>r,C2NfOjpcm{Ll9vQOT>9U;])>6JdbXԠ `Z#_+D[7IIjJɟUh ҙ"`"a ߒ"G̾H`6yiCk(OA/$ ^%K^+(Vr[RR1"u4A.1X0=7f/"(o9/L1X{]q`Ȝ/; 9a>E)XOS K9mUxBa"'4T[Jl /K/9,rlCAj_TiǘP,:4F%_0E5IE'rX-|_W8ʐ/=ӹjhO%>| :S Px„*3_y.g9| ;b`w NtZtc> ײ1KĴ{3Gl& KT1ZWX8?C]~We$9; -.D087?1a@P5B,c}jcGȱ WW/ @a#LA4.ٹ^XڋXٝ:^Izq. ٽƎDn6ٹBc5Lt;3#i3RAٽ9| cbpcTfp> 6L/_x 'ۙz7~w~);qU9GDT! 6]c_:VlnEUdn6UˇKU;V`JUݵޙEO[)ܶCy*8¢/[cչjx&? ՃJȚ9!j[~[' "ssTV2i sLq>z@JM->=@NỲ\쀜*/) ̞r21.y? bO]3?C!yw3ޯL_Su>o>&lrw&i"< :]_<<7U_~z5є/rfn͝MLmc 6&)e+n7cyy{_~궼07R7wPuqpqo{ߟ+[w_uOq?u-|?WS_tOq?Eu-L_p?Cz .e ϿO*3 `Ђ6a-`kIf-s,RL-R`1eL~dپ&+IhYRczr?㐟,v~,b6)up)3K,RLW"Qd9JgT\1f3@Kh% a4x,kA k ^d kYj5Ah𚄓vXZhX1xҖ51Y +Id ZZ\C| fD>hB֡#-$+Jpሟ,Cg:6 3 xH "}C[`ӨOAFn5ʬLHϰ:N@VcyBI#Dr. "h hg ۃm-qu>V&൘ G7qi#^tҒ[JI!{q*lrD܇Gk@;oI<5xZ4xM"؇'k!>V|lk'{d+ :sXӄc)?W`*|\v aVT0"tMًcΒVz]T.C$cEp._0M`AlF̤@U' u,—rw=3}resLV&ԙy=Ejl1#XX۾;R;+[$4pjfљ lݍ3)`xvcZRT\%fNV Q)nsX }plMa~;Wi+f{v%Ζ/K 8WPll{f_WJ|8(A ä>nl"jF;/-R9~ {^'##AA:s`uih F% [U۴"qkjXS~+(f?TT)*qy+QR"tJ8۷)'3J1>pnVGITq3J&J0CQ v&P_񾅶X/)T/ϧ+GJzApU]<:Yn\~%&58IS)`0効<9ViCbw!bX%E+o*ƾtNU*v-zߞϢ +4 {e6J697@28MZXc Ub+A_Aܲ'SoO1ۀS`*f'r[8ݝYvjҩJ;}]|Bޙǖߔ 3\ a-`slԵ怕e7ːزoW|A\Qu&'9~ l|`pΕ [Q =r#vQu0 M.1%]vRat'IIc(Irw~Z"+A<sX4*X FVGA<^^7 vq&EwQű:؁6y\QbR9GuB/S5^fa;N(hz)}_vq@nu@$_DVH|08W12e_ʿd{xlzUܝlNDU j>zƖݗ&!jC`@ qэ-V Rt2m%K6dX)"]lj齔{oY:8VmS!:Wh#O0} :OVGL.xllT_oqqqLec2p;Ndck[ Rh6T#0H Q}ppS@ώ@#gƖ8sѹ e^ CZLu+."T#yrHhlكʼE-X'I^=bKߙԘ1"+< gb`[c1髰?(o$[eR6uOœ-m~)-&>883\6y 8V -qrG]~.3jsqY~ sjZ+9[rAJsT=~#02ݬf¸9Xe>sY~ ae9} x* zjC.5Wg󵸊y!1U:pU!ƔCm-7^w]斻~[hW$k sE0ڊSq:+EKٕ|dvvjjy6 æ/ML-yz,ZlQ^oAn-})xǺǍ--qcl:WLg ӁvJ[ǧc~Of+8qpçco#rCtKӫce0!Y-+cxMK-H_2:Uu*corD~@N`#m~R:ߙ歼!IZ5>H;0ޤ:\Tq]_\_>e˲̿[pWEqws]]|/ǫ\}/J.MLmc ԗWrU}/Ǜ+sYn[ﯾeywyY]]¨Kpx c./mo;ߟRy*4݀wm&8֨Or4 &+Bs=8'kP 3 |}44S8UXi;f;VE7e4AdX-fS烠1Uܦ$lznlq"җ^s RTn|RKm;ԻZ3)`S!9| ?}m*2@"G{yZ${˪A6yq>Elq*E< NX9@: Ih~|Y4sopp|v1f2춓t$]mSȲ+./TQu y9 w9:Ehb`dc0ƭ$[<3VBfFܤQ6i|ד 5f$WžZwZ}5ٍ4~_m ARߑeQ{*mߌcm쫡T2r YdfoZ[èY7eUϛ;]YsYze<*<}5:[Z5t.`_b49~ N{kO+b_㚗7$H}5fZED&ޝL!j.8N=Q}~xa k\TӽWi.?"OIU@wKʰp.)qãa^4"懶!KnrǓ˒ʳ dmMdR3+x,ϒnX/K*n&%;\ؾwfw n¸Fl{m K(TYOBӏ9A#~.tU\1OmF9h2 d` ޓɉ_Ldl;]9!w"}AvxGi;Dx<KO7uRe8YQ d)-diBLvn]\KXg6B x¹?RGߝ9+& 0ё7d?2~hOc۴ Ӿ ӿ0 ;}jUd1/V^ C`H1R>~Ӯwmv?p_J%t߰y'x/}dpDo?2m1 9B.R6grPY}|uo?U~GRfu]T'lW(^MjV]Vv`,|K{]r}nvF#G#zm[֟2\8AGHd^R,Rϸ:n9\?KV. b #2:qezbx V;9{OߜSxuD&uV ?7gG\OYr&oRM%NhoM紻 W!υk/i"xTݺD$r&弘R{\+^;dz&<7N7bZ2ۙ7eZ/0¢"([OyXkHW:ͧǬ`qx.T57lN%b [4lE !˦ Qq̭|F=&qpM/g4*O_>}1' )|vhLY[5;.Xgxk տؔ4l{8=ޣ^#~z= I,UtVWuwݥ"GOiI0l.J_܉ }i J9+"%Ŷ~Z=>?(fGp-^giKA/؄jn8{]4OlLXU˄B9'6X\&9mc?`>0dY=;I~9L!#vi/ D1&WNd.Q5pg?X)2Fb{BgVS6.q,^Y pR'X-!y>>ic?`d6EC*OA)X%W'.s[9gxj͌mZ߫ PpsnNm( mG>tK9{,:}-5DpTޟwKNe]۟)yL(]Ṵ c'g{U~I˺aYϺ [!~`p]J_`÷V]0hb\{,Y̖.FjG;BJ"Ti9[p Lx]|Gĸ?δ(#y~T.%SP'LDX><%Fbd<,0I1IRƕgJ]$,* m<x`͊$L чGi+o+=*<*|3ܞMPXiC b=}x0)J'&x;,j>O]Fe=xJUï17vgā!{D6*+|eJv4§og6';<^&DKp͍@K" Teo1Rˁi1Rpx_`gD4_4aomUghiѕh ӣ_&'26ayRdig_ftX/4gbV}}ʕH(ؔvv1sq+X}1:<927EDj'G"!oCBos")eU0#rfD #f̊6lP=p41*/'nN @\S%ڜhN 3ϻ#57]7C5T]=m ֐G95k Kx\cy5$ z Bt ʘ1" g#p6B`>FC:D\9mljAY) +AoW *4MWLXՇT9ݫa w.x :Љ6@ Ԟ%QT]Zz_:c*.7]X nJi%˫^/;2z4̾ {LDuZ .r-bY]vtR2]{L{G D`NX55tEXanɫqf]$wf[0Q685Rcj."S3j*y)e]6`TwKKw ]op).S1pQg翍8e.E7Ql7ә̊!&jh&ֆ9`ut/z=s4GIWuc4 Z8An#ßrcNgF=y;S!kqMp=ߖifZuۇ\rzu ^$,Mm=VCeQnǓF}SZ;0YbvW}↲!/zhhKH` >$cz.]A6IuTm\j\U 8+O#TOhb8 Oᒾo @=pZnMzz$kz,U4P]Aq_oS)n'hC #V 7Pt-e 9mս\+!ͻkNJpsVSMVu E[p"$y5 x&̨^, 7\z,1-A^?JeYzl-<1`4Ѓ YD(Bh#y7ȆEnm^h,t PW{Ɵh{ .n=jS[~8llG[] nQୟ؜ T3:[jUa2ڑbn ^m jZZT-J͍ 2h`qGo>yj>-1r0= c8:p"ez4xwA"W"̆ >?_>N<-;pVܾPq]|%N+)G0yBoqߚOUV^cvB QE ^&hB*=oTlQjt%}.?yQo˭Q8D,VmR8ʽoܱɺ:ѭ&`7fO,QZrpӃ>;tx~]tpˋ[cjh:cOKvvD[-kܶki4O @ @Y;Wzz#vh؀G8N8Y5EKEI,h'66$漹pӒ ѕݮшoh64N8a@oY$a>- DDM "e' kڭqz*N//mwPӚ@RN#t'JnoBhn4BBhLfp[ @ %QBBxH\e z! O#Las[{ xo#i~P^-MxvI44 #t25k)< pqv sp(tq(0e^ x3o1 s[yTC[nϣߛF4c1jMptٜk$) p.zä+P2~$0})w:wE1w C sd a{0 eaIJkV0't_96AF쯤F 0`]uD|V3,꽖vI Bn#IFY:"JL&L >xdtw0ץ<X 7i*8y5RK!2y7I%+:$QpK mv J9(O#f-1c1Q jy"]CÐcvd N2lp9̑1H"7ʑti؛:A7̦#CXf'b [~0`k1όϊ wQJD}և&l-PMԧܲeUq\!Z 5?PЯߎjRˋP4)%c*{h2I(ȧHEbMspe=Z_X 7+Xe-+X`t->6{ *uJam-V* J-h?՝eLzMcَS,)\`| <@|ԫAcZvmv] #]4;9f ?@pGltKc+Ű.S8ɵc C! (3'om1Ep\ݠEɟ%q|!==RIClcXWghO5IwKރoUq&(в fzvwM'p8$R0 -fpL:8m@Y7 <6y'vccYaL.Y RsQWO1Z*gb_P:H=ہ}`O\ƨk²!@(,f1*=\)+o[rwC.MЌ5oqM*`kFzZWo Mah#k(r\ZCM~։pDf;§DF.Ț|fSx~Q{6=߶}`g۴a+£GMs?l~ ]{8k[>,NӖ,>>Ϟkf/޲ܖ;1L5Az.}һq6@dC?W]);= /2xvUºj袑[LF Xw&N!x3h`m4~e3[WE٫1VDE9!@[ºAT,˯el"-ʶ mO(ۍPBB6.u)ۦO :ثʶݪg_ҵ;w-w#R"jFBB- ڞPk7BjmAHڻjP{ BEB-u'ٍP煄:[.nA=n/$݂PH؂P=b7B [-mA=nz/$[C.mbRz!Pqz?&ҥ -^:?Q,J,z^d0zK4Ch5}2MgSޤh=ɕF$`<YLЌTt5fC蓧_R6,]fuinצ]ލ޵}εc4Hq)/˟CL10ɒ<&Ane>]gf۔3 Jpd줺r(\8mӫ)&׳i2V\*_!5|20u24DH_4:å߿havh/*H,i:-&f4:KC2J@Ys*qJ0w{ ˬZ{{_J:gXixKEEDa5YIK Ȟz{b,8l]2 eG|ͺCt~CY|;G Tp8wOf(Yʧ[.[9'+A1w2S>suPܲ+G(kFᙾ@ᘢv!` Ƥ$RTewj 0@L S>,P s"Szr|î-s`L֜) kԞ7ZU.'r *~À(YJY H VہjO|TD&k%{؇NIk2uMֵ;V7 &FA-t:Y:t ~ۈRV{tm( GOy]d`pNPi04 , Ŝ.$#R`, T )\ _更W"{Q9uS 3![?Xh>y0rCFc!8Hat+0#X"u(o<}"dgT3yg_r`YKHf=ȷﻊRj'+|ut[]p@ˌn$8f&ݱw>EG+ṙ}Q|~18q-hGOzB 0Ա;Z_?,zj1vڈn|ڑZh6 S 0a" lCU{6Aš v P=HNd)HR 5VnWͩן1.h h\Rrz0m@r܎ ڕ 3k?7yrX+7*LiǠ͏C/Nu2n<y. 3]=\5W<㫫˅w*jxts* A;,dx[%gq$ Rhh-9䟓%Yt"[|p}{2 HV8yWkҨS=ԉ+LDr0L(`rΎ9 ʃ+9u^+ZX..e,xwA\:E!-dR}@S`0`ckeHװsF0&!Z[ձZuX:# 1P$)N ǝk{\mʍ$)ĄL'7I<+FK?cgCz ^gY;Wwu9zH,F˽'s'C#^dJW`Ƕ ?gWrHOeUkdfR6 4{$`^{8w'U^z!l!}tVO /1C )F^PB)e`ź8W +W흌4jW 9O$G-vJ˯>` o/#?[ rX.{.T Y/?I:~QjsIߠT6gk> Tr>ۓa=Io챽8K܇;9I"y<ƍzZYbFuVXބm۵s (u"[eU_ Y5[D?k@4iV+SsP_e#J9cUf5QO"H@HJɺۄCN\FrnIf93v k`37ake;je7K?:}O @5Y?~M`z??fۢ@h_7̞}z?-mt815?6?μ+³Qis=m9AkMt?r S>?f%:=2Iis]Wmz%8a B9ޏ`٤k1NtXƿue}郃˱1Xn+C|ʭ?7,t?#SZu_,yDiA A%!Z x9V?PXC#|k]ifDٿc>v͏2b" JA gU3 ?Z_(a  ;G\,UWPQx%r!fFK8Nq'xe'δ 7:ՔOq|B1/ ;2 l.LȪ[vu67W3PIFsBf3ڗwό}cDf!f@qK\Vf1(f`#l0pXҍ̗0B@1@1k>:bp9 GK~ ~lR߼ @̀I Gbt2=|Ե`80&'GGUhG)n}3]r(x>+E 9,ME{).^m^8B18m+/ b!ʚ9 ^OÑb/w]pnl@*-0KH sL!J{)1%χ{@8f=p|l-XdY f sjK6:chb<]^Y7mH!ȳC+ B)x\xA, .p| J1|+)# &˹`އPq$X1}[4&[(e ~K*a|/CB`DHtARg;LG1.7M)WX{ ~d$6x(.p|w ~l#Ԝxnlo W. 1u!.3ZfЯ_Tc2e&p;:hǑ1Ы`V,){αI}W9VvB0*>MpbG3vЕNj+6p*5kd-i8R,Ayzv[NɱpU 3JZ@%B6qI ).r;<]pr@(2 J QhBx}布 tSdւȅ=a14dQ>F% L&X8*=p46F1QsP+ǧ2AT=6^qo4WHԱohئ$ ސbWO=sll)}. fqp^Kxn43G!GWef}NLֳ*p Xq \{M!z=A^5 n9Y2اzymZ ]`vQYAB^Zͷ (f²;^%ŤA >وOk(sN3i,Dآ2 +3bֱ=I1VX6PgAaAZ%$DLD åI1ZBffshnm-ǹ^]*Q 1%wCq)5L *'|psڑ+$M>̃{5#ɫstkqNQFq'i;2E/X҆=,DS"t_D^8;7+yRP5 :9-S,uw81v ^]IݙL<Kg`hբp<:A'^nЏu<Ϋ |l3S_RxUrntH> *aID bef14#vu:REMΝ"ixY_REgԠ`̵#HBټ)zO|c8d~3s.?ϊ$GؾfY-=a쭯Zӿ:|+Cb(c]hcŲAay#e͆(ޒ kTTx)!=n+H#iw!ge!!%Q{5+h\;|ZGpݰS37G+uG}ƑtMeS`}S* E+36(X`v[]p(d0`KCp.TQqU }cԡ(,]nP߳/a?\"ۃU\ &\I7TT~nj闙 e?#ahE =cH[Srp(Nʖ<֜1LتM-& qrbrrC<:G|)U|,#T-{_%P$Oy͛rѫʹbrɰZ_8XbZhИ[NOZktjQ@y]T=_6Lȹ/ݽl{hovw>MHj~ Dmᜮ)ﲑ5x2g|Mh8˗.D)h Rnpq̢ʶ1 - *:bq ;[^pF󀮵nQN.Ƌ[N{>'+);) Qi}{` rgPE}3H.n=}킣~;ub";wI_/Z֫.O{&ͤO J DPcLF:brUnׅӁ/z hkWV QVV {/nVs2(FghnuSE0SDEŷgGHZ|{P[]p<:bp9 G.}Kz EAn#6m[x?v]fQf6~^m Aɳ=}~I?5}{ o9.L+:${kl1: _'\l/Eat/ʘ? =@QоYf^{Y7zP.>GC}3]pzẟAmo絮9uh54+)CKdK:!en{z߇mVe^`\*D(<4K**4f`2*:CTW jLnvg ]yF-U.p: <(OQ[e/V'C_V)z,"@I9OjD2Yo(N>Np7f+B,׷.8n^["&IdVx1ԫ:уKj '.AYf,\xI_/|خC*x&{`i3j3|jbiuU.TҮ:+d7+rpP}) f GO |2fjdTq ֜H]Xy9Ǿ'޴b7aX A'DDjHLB& qb"zwfPȕK,|Y3*0ERs:Gp#a#`&a~^j8Ƥ>CO{;]O#GWZ67Y E07wt#d؉P=oͺ홰nƌm/̀Us;zYA1J\iRsåpX|=ML~=P0J=%x\+;ODQݩ{ju9XQO> /rc#׳2f]/`RD@Ͽ8)e4@QMUӛR߳00M=emns0/Onszll?h?Gk0JۺDDIq^S ?$ `dN> KuY3Q\R^bpI)N8ޠzf-u|mr461a4:3.h0aNqd84̶0>;O*σ/o3x.# v NJ:?_VpU-N߳?TT{.G GWaTE)i^, 8'|Ko(@x|]%B?I*c4}PiߟD *pb8qadcwQu9TKV6~ `wI o=)6{#${D̯K1>{[ml/dlf. 9UՀDT}۝W1N:K:IBH 0Ht; ( 33wyC1JLbG> >dW@N)zy҃{#'1ŇIΰo ~?<͋?Ȋ(4W߳w eѺɜ&sٯAS 9 Ь>{A7(O:FEqWƗW᪌ B$4!i8*z~K8 +L%W=28+UѨ;];z:X«}3yy!3jrDA'dSN٢FiGë[7[c{eT!ek.iMZPViH1YrΚ\,F3oƦB9yʊ dǦ^nۈ9vJu|u(\>G?nX$Hv熜J¼~r 4MMCMP4c  Yzfa9SqX?'8FVY /U9/[+1Sv̮wɌ=QFq>UM#Aw‡"^0\}Lz=駵7ٷSwޛ!Gj˄SuiF;om\fD&?*p!强2S/1Dɵ)9ζMRҼ2TgُJL@_eWrx4̓5k}_rZqTXVX˟=/H;z{7!ZE 1ʢv&JNfmyR`ZNm :d~fV138i*ά0oڳ3T.a>Tn1ejk:v` Jxh\ Oz@FW VH)n<&ѬKu$v'ioZ}Pp_GVYr{I9FŨ6cxvdIgb]R*Zj)AbDds7,ׁ%&7Ip~ <޼K~ ԓMice:5)7[J/Ӄc{#' {)^ߕ\ wU8។xS*iRCg6"4ytBji`˜VJvot;/^SȪ*dK mCo7{|>ؑY"V;GJQ& \Ni&Mh;M픯EV5HQ; 6+hd{IH#;Ӯ-cfYfG)^"Dlw%w 0^P!g8Op_tuƤ+5s韃V0m3+1:!ء/ ]-FxvqJ'|cݢ„W`f5"D"< _KxdEdW0@IVU z`<R:~,nP}`r=@5OK+ĢJ'BpZE /" jR0`_]d[<MIkX-L*&iR;`_,RmvvQ4M/oת&?{W3"ܗ75sb9"J_|JȓɆ`tKڡQہE:Nev{](&/GC}<%V?u4"leLLJey5iBv-x B|So/SlK`W?C-]:7 ̻sJsWv$Vq9.NoQ_-Zbt7Кׯn%,rב Dۮ kPkgm_gv" .Dސ]qަ~sQVRta]NF0 =XM\)ACF0&S̨BQ7A 4 W葀Cu28F_ó%M;*oM3Fa<' k\YݵziIOo-(Dx0(n0j)! G,>jC G1"*Qb6NFv,k)∐Vraa;jvh6e0uk,cUU e@ bDINaEWƝ)#v("AmM;jdM;epk-cZ2!RPpYT(1ҚRt_6]2v2v2X5T Ϥp! yc:U@؃NB4 0BI&M;jdM;_2.X$ <=Y-;y"%f@LC-8y3վlL}',Fc:nY?!j )a~Q> (>=a MQ}!?J:Cן$E; !{_7:| j-L=sbJú4*#m_VA֫qhf'3*so Rob/JR(-bbv>Z|jl ޖIAۺ"?O.r%RyY2 )Fx] /fu*p,HT2ޏ-/@ƩdéQ[47px}{۟;R2VJ?=W}K2<6P-ON'M/ UB-fE|Uul;7R3 WpEu+b{m<),Zђ( R=KbvE-hja4aq8-͠y~ bAS<~)9jX<9_) MZKפRmf"HK)jdq4[Jn'tQ!=UFj$57m8`VUvy|΄]]]id4O *1 UuqCDmoc SVDb9鐧^:ϭcR=~h 5z};Z0CDƺ3EKdQ)d c\q fx](k'1sӻd8<4z)B)U8jhyn%dcJԘw ,ō. 0@<\)p F9m4C(`=< Ur.9~ɞk:zu)xݼz݈^sEF);qunV5N|޷U3 u3f zHtfb*v~Q 鞆&w0‭G*mj2>(9DQ~ H`k3qhmuOZ*jGGB` 0 *p%*`DbH<4y OIlPw7z})uGɍkBT!OG6$;ܦ֫`{;PVC!Z(cs,e)E.P\0ax{ *xB~ňz6G@P(EAT~bXОN"5ɀ9t-KʈI X(ARN9RDZrN#SъcbB  HVB.#RG }zsi(y=ކ4a+yiP X.Bo) Q)6ɘ/Y3[X&2&0dcP[+@o c`V[#a!|\wmY`aˮnzwƵvj&J@#Қ9$DEh3m`>&z) όɴqlI) YY#y6l[8Yk׀2M{9S(*L *cIYe)IkY-wØtcwe4gQ?@\2a+ȏ `Ag؜נpU8hTLW}NZvc4lKShiBqBB2ȯΉ$ISvl0@&(h}~ _ _E&]Gp̗RʺVB`2*_n\1D0~R3;s#UbU '9M4ppܕ ηNC! :Sfax;NW)=T@P煂1R4 ?R4d/^{GEڄ_} ,,uܦa֍ݫzǹh ɻ򅵝Y6UýTG','akE/]󡮞a}&i1]xjo}B 7);^dHśRvqQNqzIO8レыrǼܩ?Bzo~zoGMDs%Zf^%j^s/n QAeJfgzQ]ewAL?&r@?Uߎ^i֥?c3$xv)HD>^O_Gqs5ٟv/?<|vi.sX}߫[~z}7љZ蟽_?{G^Ͼv/Dwg-v/>F)zʹ?{/8:r_kƓiA=/:>7> T K-tO>9SZ>[g~3 o|mgy=|G{J7c}|-:毽waPd |콢evXs5J tG*w'fΧ$`9a_:ppw~u{g_ ʽz) <1ׇ4u7'D6krZoo7w[4AӒ?yP뢃esnc]ҧ{A}OTFBfrRhBLSfrD4ohмplhPtC&'^0/M"Rh)l:|_!6麐Jr&ِJkSJ;KCBcgO;GZg>^#uQEqin^+ɧeY??,L"hlvv|We(}HμئV^^tziSYšz?[3daf+N+{\vV}Iφ4BP2u;apE%*ܡDMрaխ2BV,xN r[}ŊIJ׿xy{"Қ !f!lK-7 9pG8֕_2sI|e-Cv* M]7C`2 P٤0 Apj^c AKV5&UqQumטwIWL!i-& 0:(|:f,G:J(UU']meoCeܘM4U2\]֜!`B V: q(pc񒝎Єr X?@ w P;E'~s 45XmUoT/7NR,tzz9 Y70fT)UmEo#E/]iM)T=~nn?-3S:bi( ރ)K9M'Fkcd"MMoG_\wo$ndtu ]tyk|֌uS|yt}B|' '*VJq eKs&v:6O*R֬Fb¶r:&*uJ&l\vr=q$^3]Ǐb喊IJt+m*TY3l7yvqt/]?`[CƠW&ynocR ZPm-t Dj2 ukg^]TĽa($COPX#J0)P *N٣N õSvHl yj#fGr:vDgl{7Xi' q׻.'f6VUM?mP?DmO}ve#՛9LI_+lQ.cm.c䏺'N̿#Bq)m xMh\3h\#3F.t\c?<0.K\c *[k"PF(kQŗhy|&F0j∢;I:Go4LڌFt鸊~'3f0rظ_t& 97j?䲠w[.K )E#x9Njlw9+c Sv#/\mx&Yg۬킅vĘጃ__޲I #dƹ-T8R\1jm6k90+B)*Ä"sy-l18lh4WA?%}!Hb-*dW.@BȆsFPՠr V9Ho Ҝ^ T E!79,J9L4F"v#2s1ZBD!!~ 8 gCp'UE0 da.G E鍄=J8KdR%\pX\߉B)9YS*DF#aW8_ <ϒԯjstA< L`UwzL0Li))Ho"`Q9)DK`C]HU5r%9V #dZ*1Qһ$ӥu˵OMS:Ho$Mf,THcskޡ 3'HF&E)*Ͱh8-eY_~9mJb%<'p ϫ| ,1:+F-\Ǵ\nNB!- ɵkxǼ F 9V¦)v]LJ yҢ񓆌A鏠X2͵N1 <Њ~ʜV#`LVH^F2eנ5kB`?R77 y.w#eKa҃B Fi4,d2.Hol}ZtCmrG7<Zʧ~MDpnB'dmM}Xl(e")1IOBtRqw8[/NٮgԽͦmrojn1,K8eFj@%ͿW%ANĄ2pgtd5>}Xfbظ#7IE\qf@ZX<̳ic!uuZۏ@!O;\W̕(OG 2dm֬7d~ƓrWt?{LjH#LVV1KM˺V1&)Uh-t!PZƖa*FZu`XTZ&0y7&B&1`c   )x o]v=4﯇%V%ϤJ[W͔y1-)<'_>D :B7m=z{7~ᷢ0'wCz*-r5'm!<_y^Q|1F|4 k|kVs;|2\ zxM-VojMy}30Hˣ^fe t~z9%\cŮk-?9٘|[-/Y^>+Mo9me[Ͱ-Q6I@6˽?Jp7wrv#:lo7]Oζd뵇 ?l_ 灢#Ό3X:u* otktTz_ 48@rR)|3c`F>S20-'{ XlT>O2p|dpъ*.HzmK̽`&PNX[ҽHulM\~D>K,[d 1%$bMaF7\Ps~'̿lZx%u{k oWՁ2λ(MܕbzUo0J ~~[owd>̖?e'?cOfϷ=DXmcۖwơ^@v HYb&=&cYh8[@)1HJ5R م9.kVu۲e/_6cp{ּ6}?Q<=QLj:kTu731Xd @>=V6+ +sbUι|b]^ٸ3/[/۝ r|z0}X?P:_cWAiѰ4}}6>b>Or:Ŵ-`ׯ1˸|a~U7\Wnj05ܨ9tl%} Yj?d$¢ +5{xx9e| ?>4gCoϡ7\ʌ*] 7~stZxO. !f1<^љ/otNWCU3ar H9 +/R+rjˁsrHJkmXEXC~T bon.8ƍ $0d=35zPr^CS֩ B c-v}w/7~#i|aJ]G?uscuk<*&nу-թ+Nj&G`lRRrZT0E\i=)Ƨol,qaN)MqJM\1> >`I<1>%QV*21Y~jČŏ}ZRݱrz6CX {Eś1HC2;jX9i` "µ4=i 2`jü |aW]ە/FVP8S}D`*j+|15}.md-!IPêFV9Q}xMDZ`*䴩Ӧ6MKZnmjSbկWھÄ[o5lQҴ72:+N5H̐};?5e+.\v"2-[ZY";f;Xj-&ZjIJ9L`Cs6̔8Q;񺡶rpv!>3=XdLtČ2vڳtzdaI4׮jkrSM0e&MLikAKd-[0+4 $W#Bb3V֫mwVo|Ҋ2LpQ"tlYΉZ`j),05vrFveWm0όt"?PԵ&if>]3zMiقV]SF fnF XO?YL fh0 M Nh^^>*ۻbBN\[OZ [=~ЏܠW+4g~ʩAbgZ2誜o O=$DF L"5y"E`&w|)9rj,K$>Y1v/XkkOYUvN2Gxef0w~Zꍷ`Zn 0,M#\*_`Ӭv8{&Mا'Ges"GnYJcIK\k;Ȝ=m O#Sˤ;ˇU>[L-!KlĜV]9cF}^K;v@cM^8뤚)xMx6vo$VYh l=%JiwS3H42mBfYighd6vA#q;݀˽h<Ւ\=mWs{j 1%I./[!Ss&#I2lkĴD)SP lj4 *mlke'ǿzcKNzò *șwFd-3E!"2_qq69<[N ʝ1ϖ+6N+VLZ{F#׊[+vnZqC[sт]-)\YpJb;|9M=LޛPK6\&S17y6sͦabmp&fVx5j7۽>=42L+J7s6H-v¨miU`tf< zƦ |&gh;xʬ=ᩗ{}ɑobb~FnPϮO׊1#p^xJ̠լ\x^rڟem0|';7d|t}!$ 4eQ 4]_\~fUU]uex gEܐW#55U6qXfLC\M2ۉkUlD&V5=j'~^c5TQ{oMϨ:~9pWw&0W}[tA_Afч\r'ؑB_JR.)uF49^HXg8׽̿< JwuSM[`]y]y\7턑TP|z-019y)Fv*p>:Y&(v*5J~`gW/n~#3f}Nmr{ y^{A 9]:G&_Tq.vzw 펷Za*_Q+s@;;rpu*U ۿMczo&WD?oB7ߝ~wQwg~W#/_ P^s}3$>O+A}NE>{]G}zOW t{x>|͟鏷 |x~zWGxj>Mu!O[l&ileW~IAޝPBӎ _?UU?o^54:_vI݇{[ 7"^ο9BUw}|]BS4)_k)R^,o .F .{Aʖ̵+D:8Ғ̃Y!QJ %!$Eĩ} EJH4Rq_P:D&̡υ.Q䭕Mrh'D,DKr&,u592,w9' )3'\əeIܗNm BV:'UŇTtb BjSjf(2 s%ʆʜ|f81 ՒPw]Z!in,Bhm,*jE))HeJ 3m ,he&G]JN ̓BKA5̆*F"lI$ /x] 04y  11kJtx2m*2t`62Bb%Rgc^i?7g>MfXJ%jn`]XI6rQg c"vB ˡGQ3\I^@|6ٕ;`MD0jik j|ӣ e#Kpqصl:{  3%Yds  @QUPբ'P~4$c;*gDH!(Ʀr")5Z_ ղ0geNCHZITf92͠A(RIïފ{)p͂ rS5/Vp5,A2é Mi8hkw~MoΎ/lyY1re* D8fV"ip(tmLaF>Pp(r2wkSa֘k>HrDq%RV .Ș4{<0[fopPaRD6PmJ F ~ h:)r]gd*DE(eBAJ@R9$diOW,XoYafP@>Ǭ7`E^QH"VR2P$UU;[7H{}]^v 2ZSLB).2jR%:9a泞+~x.LmDR1uBΩ X;uM4 B4ֶf/Uw&H d2 D 9k듵uN[jKLY#VFR"&|甑plҨ3BECt]( Wrl$diQΛU~Z+򎊷!-'r^e[ޯMjp7sեqurƎqbS~ٴO?=K<c=>a뤐4b_Y_񲠪}ո]_|{}s.qyˊs>`u}[۠gݴo}9-R\;?Xø w^[/׍r嚹1.fKbGgN|d>IKiz g:]n^z%;#_&goXޮTq}}'b@; =$LOz(֪= /_P@iC]C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =CPLWz ?J׍ h5}z($;Pߢ*rj衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衾i=TVvЍ pнC1J#CMK9C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C}[zgNyq%/{D`#+xŢ-Zn`~\';bz$X45vX 2-mĶvGDL& XHT'`eA'`"XT5 VwJ-彊;B/ڌTu\oo&_^\DRÛ'r.>b~zPVۉř\lP^8[D"{8Vi},im/h-:^|i#Q/q6hKD֟m])x X.zu75Ha W˩vy}̧fqWez ?ٻzFLoY}+./ͫҊ϶& Pk XCJh XK`DY$/qupS"}`ˆwUb ;e-%VN^\ҋe-y{_bYO^,@6b|܉0"JzN*#INj# `{1tւҋ:i; ֲrG׊wV Xk{6>( X2D,"qaV25؋z+MgEvNUNJ^;,IIH/;Kָ55h{N^:NzE`AcvI]'`S.i/{K_+7MR$$9azrQײwۋ0gWO_| Lc>lˬBe֋*yY/D F{?nNŤ],SL?wGV E䚥Mv!x4ȐS-MUr9\-/,ܴђzcp2q7oϮ'2~;(N=KlOk~{\/򌅭GN_ǛY.#? VKN=1ȝ:Nz4wWO/7O<ϝ Ց>pm;XE'`yB X! - vKxwNJ8jR/̸L/^s^ :8E` ֜/uNwVuVq X»N7RbYkd'`ZN:Xjⳁ;y#X ;ewVr- XyW^vD;{cěo?M‹+&ep>웄_c|P`l)2U-Zo?eߧGݨ~_IWq|A_pFCw?vNH_m,~-ŏ~옫ERbkʧw;e|O 8׸W[zynzt [e+~)}v Rz!.;֫Su3w@?c?gч1r?%B|9joUZK4JP! wVfkQw O^;}?޷U-,P5[Uߒnr"h*kK�Xj/b"%$'aBnE,B[ RQh]jԜ(6Ͻ3 7E%#uM:jR& Bpg88gq5cI$4TK`̶f~BvkPk(T2#H*nפnEƖb1d,Z4ԑjΗZ3&u),N`R: JU$\Kc;s5lp3uʪ)C)Fa]q?x3S rY;eeq[$%wrwah&M;JԸ3[~ާ\1ИUFҚw-"J& 6v .Xx!X<>3h_uuus}|;Th)Iu!dHXlƿ!erZv$Ϫ,U%X3F"ΉT,m=DA2{gsVoKin \!Dk099į#lf<ٮ6Q.)ss@"P+Q49KL }ҚVeMŁ ͂|VM<⢗6$%UVش+!KFL.9 /6v;4;iۺ\6=VR]J y; TJD|-ˠV=!r}ٸ xH v+ ,*T~k^-\ ԃshQ3o^RT˃laSbAaenX"\s lmC8QֽFf\0-a6p{(XG3=U[>0Uʳ 㐔ٰՆ,Zex@ƍ4dkj][Ty/΋^[5 q X-t*x$DX$AHeBE+Z#IvF: t꫐Q64~ , _tk5aUE$YR{Hu Lk[0N= S57]T? !W0c)(l5!^j @ mld%L8֔C98Ƈ0\"0eW&S%sJ vfS.JA`kmȲEZC@$;X;`30kj1E*$%T7E5%R֣B&ݧoߺ:u%-n7LTAjoM@ szI%R92rNj ^ImH(@k/u4)&((H5㲩G*֚}9W{e] %AO8PGy>oIfIB¾ vvRD ɝ {H7SQI4" 3L9^TMJ(4n7Ai7<٠[][/I$6Ϙ"9qp\1: l$djD) !ĄeJ{0|g?`kc*L.u-x01uh.Dܦ6Q # ?冎 I!ԁgND iM0d̳p+|bmG@ ;=!p1%QHv 9A+5/+X~Tn˙w҈H>DN2 ;'8Ltv7Xy'x1ka>(svC6<oŭ N.tLt*#3PÒbIӈpۄj2 P $2 b|]kU/[ ODjjc)VZaҘ0;KY5EuԆHTDPw@֛@65P P|3E`ۆ\z`shc[ A )HGP!dуlIW(upZk/$j|`A H^Ic{Zd5HD'( $OzlnABQ;H("j,*fzaxZGP>zM~ R6r$ؚnia:^- siѨgYpxZA(1--`A-@MpKA@& 7,!"bHeNm-j<3kxa1Yqռ]gӴ*W&Nr @ea]Z2E42$c`]ӿs;DS]Vg,eup(EʪGCm2FePNrrӐJ3l혍9n(A/U5(YQrrbs">rfCՀLփJ`Qܽ둥%Cd8֞ V'7QA,2L$Ce 97J`4F^}EGhK1&F`p*Ѕ`F !K#r^\; (2~Uaz+/~;gkqllk|C8N"ʀUNTF[Ԝb?rZrQڡ \/yvߟBڔ̭]8F?׽̓LOɟfG+V/8&<嗣ɯ\dlęN1ezbQs[{JLq p><$/\!/9@ *UCeFv=TF1衾!=#EUPEUPEUPEUPEUPEUPEUPEUPEUPEUPEUPEUPEUPEUPEUPEUPEUPEUPEUPEUPEUPEUPEUPEm8/2$=Tio/`P@k_}(dNE $h(z*z*z*z*z*z*z*z*z*z*z*z*z*z*z*z*z*z*z*z*z*z*z*zo[jP  衚P@)YC}z=TC=TC=TC=TC=TC=TC=TC=TC=TC=TC=TC=TC=TC=TC=TC=TC=TC=TC=TC=TC=TC=TC=TC=Է-0+0ͻi^p`%+(})E>v<Z|~/F!>e#XQ^=&<մufm VXXl@z#aq3~]MfĬT-.24&@#}z NbrXc1֢5/ V  ɂ ʁX,o@rQ @!l@J"v(@YD$/~QfOjXJ8az `؊ ŲCⳂH@sQ^+ط7rL `6Wo(R;p9\:[.]8=)`}`W ff$hJ 32-R"΂=#B@RJ 2ʄe9傪2AVR,Xˇ4U g ՒہT[`),%`F1Zd(_ϊ (n \ʛ聀&ԫXJ:S=Ia9k^seޏ~p4-F󋼖96^ң_rV?Eڙr0;gi8]>ߗ|Bȣj2 X&xYtO0%^X|X4gv `B RN}@$i a`K>1/;70{VzҹeEv~ z tS.e\eo1\=::SMon rLRjգϳW5f@)f`/?f# \ |s-zRb6IǫÐWUTFz! >ďi}qXO}ǪQɼPkJT~YH(DѲ㭟V?~[}Miİ$b2w4Ie!l}Og$M-4 GMۺq4p0>?fx7ęRtCGָ/Y@˭ ^`怅c->zBNjnJB-c4x_,Jp+1d6χ8`>nk0>"00>F#z8:k+3:WZqE8$#{xw^ޔ$ipY[HۚTs_ gUe]ܼ10!!ndGhwm4CqER (̗\'ſqizYP ǽ ,-pj99 =ybOiv^f>fti zpyw'8^y?4S Adq-sC;6<cK~?\9bWSMxw-_^_À$x1Y]{Xigxy:6cs'^_VC;Xȝ0y;o~iNi>5FQp]picټ~>M??wѺk= q٢H+8xiOn\ƅ|qu q`ll:Yg<]F\- 667oqs [\1HMEQmGיxةo;UBzX/uՍQj?}gw [TqOO$O>i"YjVZjVZjKcXM$a]hmPmW7-:KhV#kU(x9si)Y :Uk>dtjQsb:$Ҙ ,QkΤ EMXG,L1N =n}̪9>Mn>Zt\2[z KLl_I ,u-pܢ*/ {v XMdHQ'nINm'Bb`N<}41pIP7RXOa=~Vhέjd`4gG*-g;a(5;:߀81;*R^-77h;RRƹt["y%Ue`$PO1&{uY*u(ùh8;oΛ^ukŹC>DM/ٞ øF h-֔f-5t0;hK!|cK 55fVgmJy'#e ѿH?NFmΏ2ޥJQ+ᵬ&uPj5JnAǓA&l߷4R{ܒcf2H dTdʮc?Юݶ}uCfi$6Y+MV僠0N }6{e}~ڱ~j]N̺pc4QFJD*<Β:تs+i;P !4N~4<$ɫDjYGUe$y;s媺}+C;PC+dY]  ~S<j*v rQՎk\|stLᇚ\g3'=1orB1P"\伢%1Վk?Ln?׿M.$QZzI_f9 ,|0|G8bjEQTIX WY9 "̂j?,4|&䊙 ~r)@;k`9I3 c.I͋"#<^!A[>rj{ǎ{ kevmޫH@uDv Dl\aF%#<^!A[>rt23d K p8.{G!vrVblJز_΋&; ,+f‰(SD5B{P)ڑ $̥#;.Xヷ)q{!0ŘsAUbI%K ;R7WYٲa*|]m^3>;p(q SOTu1N+Md9f ӏ٢Dzq_|Jh<{mDHad"P e WZS27Fox˦h7FY\\ktu//P¬ b.Gb`DlH%#8/%+=qBn6ǂ I{(ch˦%Un:Jqj(el9NnԄUajNz򞀮BԨ*WQBN ?7yL8w6emO{+6*pȻ} NОd(8cp7 VL?kR&Qy9Y ;uXD"Z<4j *'< }o^_o|WY>\.J%?f`Es$뎭 NƸ7+=qBcrx) V^ vCKMHmׂQBʤ"'#<ΩvW%xan {JR0xbs=">~=v Y:YO 2~;uu5 $MdKآ>"]M"6S'EX;b`(6r:¯b-t(XMQQqe(&X F?VZ>r]/k;9u_FOΌ!o5afy.Ix4L_7ZJMIJ<̦w|.c)>'H C.LV#QւJ 0*Mt?jgzLU4U@:1!+ ӕtc\I$ '}1PTnF o8DۜJSLaz9DŽFє>Y0[Zǘ4܌#BFx?Y>rYQv|(<šp##1өݿ+}TJF)sx#CsHEyr.6DKd +f9(9ެ6UYl+88Ir`$W֔Iܮ`Ө84T58zƖ$Ns_iۭ%ykyjrxʦe6ct9aU4ӥtrC뎈Ʀ@8 ֤d-QҷQ)yл z#نKA1Y{cNa˨C|[ֹ-[p@y>ѲĚ u$'$-YS;%%WQ]Wx4i3!ws2~P2~>kN!D\-Μ5VQaƌL#hEU SR۫6=]o{ڂ:T<.Nh "8b]RނcBy͓TN<5F_F- 9]ק(ޞCL|qowI츠޲`F5SQM̹@ײN b) nO.fS=jLP9[4)xNg+s?kОDp)۫+eTLHM#qݴZ*K£8o&:cyC,\eP@>>{{<[ђɈ&c2CCbъpjr02I9< `(2TQO<:F_C- U'wsz0+wv U>£ulcBxhY!Q4Fa %6[Dldʳ¸UGx]z9aB-wF "bF 2x!hυճԣ?Xj7|2g"J',u%h:9-Z>f#T\}=ݭfS&2Pp0$vp`LAO߮3=:bcoCA=-pЀopP ثmw$lX|a3O1/`F*JC/,`4/1eP $':i#<+hl|.QARp]qHp`ަY>݊Ֆ2L(I.Bi#e"MqKQg(m&5 5vBkoڼdWt6a4.{GN2P9RGx4"bwlD- ݘ ݕP#<Wb߰ ^MO Ǫ`+ IjQV&.rףwH3~Ӣ40mZԲDi1+ Tf8rJP`"C Z[Uߌ.3`O*>mt{;j؏~aI70)S1**\${B7`H !ZNstA]A- z,LW݋,骚# F 4 np~28w vq@U3ڽ=YLo4].Oc ϛUcq+kߑm9hM;lTG 9r)y >(WѦviB;~  OՔf{*Fg{4X/g?G$*e&~_wW#PSD7~ΧB8uΗЍw~M>~N6ѧ'*8-D]zZOf!3&⌒$ڮOv\zEFx a3QJK9t%9>5+_yθJ)nRC ?$:xK9j{X}~Ym&UʳgZ_ڡ E`ˊ vX֞@P=԰IdHDbUd{U]4$ćGD&k"+ *xJDKt\Uu%Akq@(8zu1M(J!fH[yxj0zڔ’' (8 ا~ Lm\̃꬯L _zUozscÂ{?0]-\bn?DpᇿMK*>=j[x>;zR \kɝvpsWcf[ʣX sߢ_aiя@&C0,SvWj(Z60B\"ZЈ,A+q2hiiM4s`71@ 1+sYSxd fIs6/8M4nU;SW+~۷7_YNPZ J>RĿ҇ @z\=tZJݽfgEja~ݯvYm,ڢB_a={bʵ {C1[HhaW5̑oݼ4V) QʼnM $+}L}lw3a3dZ4ӠdqXvs٣HGOZj w5HNK!uAx>`y}#,ٲ161"e&፷BrZt}L6/P=%qjft(!Ox)Su>n$% ʃQ} eGZw,ʸ][5h1Sɜ JP̈!b(.jE+r.~aEGy1C(r=K@ɢqQY5s9 #~Dr>~1ZzKe&]o~ֵԤ#3Fy#uCEEmԈ@\2pὪ,'2-YYU5L 俳G6;1PDc l-c$_nWw%P;nFinEEZsR)$bIk*W>ISRwE=-øWEBy2~Ns8ӥ^xiS0zb#hud[.ϟіzoCY0sٲ1ƹeQLOxD[ A"ڲ1 e󨂡SG Ƞ%) S:PxKRo\|ݍ!"ڲ1 Rg4J]ݖK ^OЀΕȨli<#SPʹ 闆Ę1ƸO`d99:<PCН0 :g>-*ӮkTYP ' =JYGb*B>v]j`:`U jHAs*#>t2iLIzؖuu C<nW~i#Ǔg7rcK"l@n2ڞڜǀM8+%`)ŖN xs^15nR0wEU Z6FQ \7wS>=T䊲an))#sУhE\RP.pxh^AČޅa`AT|%`v6mgExci=1cfrƓcYNȲ/O1f uk1hFo2 7aMBG{rYX' {[O8[}9e@(^hZKC견):\ŌޅBa'*jEY3Xj`wRxa+gap:3CxJOT$R 0W:h=X{{1_cQ2q@,%2D:Ԟѕwu3C'`'|(sfvh- 01m%TLYH0LJ4٣PJGJOa^4-uv1Q5RM*eߌyc0|Qzm#lWgceFdK)<'LӦJ"tNEݩ^>|fUR  Ӝd& xo ߾ ~#eq:]vTSR{H ^,e g$rO"뼬A\/cm>>Q:1lrv2fzxN(;y*7%v~6F/XlTΦ>@.חy>fSΌ!<4}p)2+K#Yꣿ/_kܽ> K !T~Թ9`u#Ce}iܗ3ndL&"G/}'{%fk ߍ$/bxD.c% DVI(Ee)Liy'|1z` 04l!*G4;'uVWz־+ڗ*1h2YS{KJΨuva\[Vac6kD@,T/ظ*9PF󇪺)ªg4gc#蘒߸it3:3T~*5/?U:{y?cf:XT7SO+ аRo ? 5f|]3]UacDuƈX@aф&KOD餭8w1q8OhD[l1SUr*aJy\>-N#Ϗh8ײ'A^gtM~a1i O8ۥ ] !]bxY4{m{Ptrbunsec4!kDT|̸hmt[l":Ѽ :]шCQG)Xyh>! Ξ](r|%OsT r6rMI(ɥ.)2DEm)wA1NŌ_'ank t tsbp;Lzb)W1:rx|lWgc䎊ûH>+-nGIAs6# '_Qc|z{ e?6֧b~&`$eKh WöO`Ku"lrs9thjp #>ՙE@i..؝ C|q9F豂Mmyr!=Qgo(Od+TWT(CB[J ASI5ѻ!lmQu܀2e^9"?ZfZTҖ:;lrɥ?vZc3qڨPPѪ׬wRߝѣP8?ɯ^rAn7q9W-2Ԋ n/`F2{0ӴIĔȮe5%2yB%G"&"nBTG dwlAkt,DU F&k"+HJGja୍,G f|DžޯO0*[c9gqEľ\}XRbt]ZbK+f̾,w3qT<A)bMfdtjjWKΞʬRj-qXQi4,CͅsUk|FRK_u+h(2|g + uvB? gþqc$)O*kWVGUʗ˘ѻ&iͧ/B:(>:{ jse1_ #,IJ|0#b!֯Cމ≞0$:'Gr&{qׄpp\a.> h{lSEw6LJ􅤹nW:Ey4]|O~~>+'onnP$sL‡W-21Jo6ujs@S/{WؑJCSv#3b &~J<*ѢH6I-A{Nq\䭺Jg^ TsWξ(F&ĵY2%4vc*J, vY2%GlV.GW7O=(5-#%m*, P`6Y)%;, l0NEj0Io?H6_')IB`2&*F u?/;OަۥΔ,>wc -x-I썡O 'נk|86AjX,JWYA[)rrk4pcEI-Qy}Seiv-BKv>>MVch,b>!VwE(/d&O<LaDx:X;?U3-3n^?f'И+*@d19ˣaHC=K;s"%>*U>/xkXo]dq~ tz2eUi)ITzo O鏇a^b#GjFv2],AZUtG!|Qw7)r/*ToEj7{zowIz}q_u.<|Ufv>_nQmx^ m^_GK;K0]E/sxW!$9A圍ӷ{joDQou4JJ4R(q (Ues|4-H6gX@syJY@oR,..\=cz=TtL!!,ǹl)5kJ\+},1;d"3P9 ҄CҕZcZu[1yPPpl rAdVY{3 y^6잍m 0o8?[+&̚u'P)UCw$^{gMG8Cl#DGR#&&0R_yP zETQ^8V55H#pxə^ rzuS<8e-cEЮ N7c|8˿g6.ިd5}룂yo su8z 7Ec7Fq#jx~E}r>Ge9+-m? xowOU0^_'g,oyFn=x2.*o-<+\I{.tBs0=e2"!.k#ߌ(^Yw-O17b *똸dף˧ʜ,G!a9 XHInvʯկoh:J`!x'gǞfc-:s^+.VrHexZ}|__^r:fWA æmvQ2Apr)(TB)Yw `g!׃g^!immLsQT`Fz.i}{TΞ?͞SQ/6laQ)bMՙ{@BghGc78;.oPr= I. y`4 B̥1R=G_[fzL.k5'ƈ^m\:"GLJ\o U$1ct6MV +s_-N۷2Ft^؄Lnj y7]G}wȊ(%E#sd'`Prh TptFepH:*2/7UNQ34*CL}]ҴƐQݏYu-R' 1ݹ.@#̍ Eea 374*yܧH=C28zF|B8Zd+A6:0 Xrhg嗗xK[ry=G?}èݥ)É+#C>.:g#A;eY*()cC xʠh^yXVlT i,8hi>pVd6*f^o቗[ݡ ;hbRQ .22`Ш U0Rx9DtP ւ)#yVL@VJ <6V~sʚߊ"{[H 23v5 fc?M֗5+]xFe| iޟeT Na_S*#QK$rCļ_].ON.J()G2V}J[Yc>CbrZ_c;d:0!917"-` `$Q]!E>El>(l-6H7H7A&DLQ<5Q}3YLN U7\Hν6B ޿2.F 50lkn{0ecuE͞_9Yv  _tܾ2*=T \Vl&0'IyK=I H*i0que4HMoJKvghT+R9e\L!)մUH'E*ҁG*DkFz(ghTG\qh92 ޾A92:귇#hfl_[ɆYEjݵ+xhlѸf?r%\E#\#rSr..9)*X;3d (?y H`1h<MIEKaّ2;e`}'aP,}+ܨŃk;z!TƉGUT/j /E7TF'#Qtn``K'0s&LUQ|/}.īm+HE]-' 7+8}q7xxQ_;TMgh$x0l@34*G[kߢ>L(]"g3N}WJQ6H򞈍 4g )>5A_,-R2hbV\@ѵ>Q&A :unӥ[Y7=Yʣtmlq9ڽ2 }$8D䙓`HIaTj.U\SqaFsf䢩o3C?JpȃTPtɃRz#%>*=mC.8B̿LQ 0Zh:ܬ]Whlw``A6]jy셢(0ך>/o90r0<$$X`9kmHʀ_.nK~0ww7OgTȑ_iI4J%aukVc ,Q2O<\>@dc~M[Z8ݖ  v8-ل1p4atoIpYrec^7N@?(M6pf "ǎ1ƸB[7s#D bC,`6)8ae!N泳6wJuI×C ,-"-1uIn1wv[Z`g V Ŭhvzf._X.^u8P\`@Ȑq>j`iT.j{( 8q[X8wdP RnmuV@@]bsG[4&ޮ9|zkTXlCa@p}'*3! 2©*b3[$1D/Fpߣu(E; sPw'~3Ue$ѫ+P65NX K~1}ӘPX wS@@Ncҹj'QVNBm$[ޝMڳ 6+.C ,S?9Kl(2ư©P6MKCZUJ䌧ʁ;9"ֱD.9l` 6]!$C>Z"066'wh1[nc`'>;[X8~2|-;E]1p*2e+ia;\} jrX9!B}`h5)izV1pL}u;U$!Dcx6T8& g5ưa&vyݎvn3򣀢|X,kN&+{z6^z\$XBh86c%>??ݥdAЖ1 ɭ~2= vɭC , eY%41;8 D"y1 6t\;fmv Pz8UYr8`@h&FK㬒Y Ϛ7to45#Z;oL{4GS8d~6{QHhǽJx$6? Lb*Kٞ-2>E9`@M<sԼm̾QY#A(!.^۲M^ o-Osk6t#u6;$fB<`p쀑 ?}~<ђ0ჿ"haźhۗwa"e嚫Z=rnS/J]vxlڎ3)JQ)s,irz<]i5=OM~4O,ټ t\{gc?c?g#ul~ ?miX 񵟢,P'ֺI\ z~uO(y zwv 5@oct!?{7+ Op9jx'rgE/,Jc_h+Ǩ}% +~>!")R)3V<ћ٢=p[>أF%Dk?=)~QJ|ve:Ԇ{3Xi*/&[_\I}R_^7F<:+=ǨIAQ|W| x7n9i>TX' _x|Ҽ{B Vk#" ?6dhy{|G/=XbCJc4pe?jN+_H+/$oZVy f]Is0HӼCLmwE-_> m3KͲaֶ -{@Xԃq2?W,s}w2+5bT AVoɼ<9ɕI15Y#a QlMS2Z_7ѿE$l"C0oX%u.ֶ0x}fYk4ҰC CF8徵\'&/1Ewi\rIژ1o9a[oPsKs\-]Z]^Z)r-,A9 WP9u*j7ҥ9-saVޘnP7o=.۽<'l6u!!g*?MBNis O&yt]䣕Fӷ^4[}K~W#rPB)q%r\AXV_W|?56= (o^i)XB<%$irqPR[ r(&-s=2fa5| >%m+zP^Pk/(B J)v8pHl5#Y31i)o_BxI5 @ 5S.qm G bD`S1`%j-h46D.%D[.{4h\naGC8Ib@^;!ԪB [3]"Dd#HXn_BxE"h %+P 3|)ӵ3fD)DC;T0x8i_@xΪ;(b 3^ќzDOLVBxV/u3 5ʙ KɄ2g-z145_BxCLmWV| 5P d$0CK(TLx^f3,G!,! 'F 3%*73_۹|)g$\$pQ\>`1b>_BxNzBZA_e/\ѤEsJ|SC' V UrM&# +!tE1Χ]u>&8?{IڑA(j i(烛\.T칚+P{`藺?x0O?MzeQ#cEJ-QB L{#8 gU:g =%jި}:6 MlTxB _PHN-"- g8uK(F:lZ$q׌*  z| u+ĭQi)幱| WbBuysb+.ɽ6IڗP`S]-#])d4WU#x/|)6gĒ;DR Q' ,P<ȸ;u2r8, i*${2K(ϙ)"RduJ6wWBxAx[fc-4:s^+Xn9J(/m]Z&`sO(*w J(`yV(X  +O^:{ 5kQc4_$Ai=CBL)PCZp˰@X#$%B( o͗P^/tC<B6lj05_Bx1ш|)3!&skOץ|*m֊2r-"ыHH24Du%j'>KJ%F*2a䈕Dc0ܥw J(T/#0l~0]JJz}c-IW + c2\>s̥Whb[#Xg'g8q~#Z?HL+`F_.uz:_,%QxRZo*G= > ͸p s~)m)דx>{_ݺtk$A>b9DbTQq MdxUo7GRgp,Y^!\YR!H.=xv_"X1\6M%|r|Q_Qx~!oa7X(I3X.u0OorInmSkmWlXS`xߠf5E;Z+{܍¬7oѨL?IȧYzk-od+|#{hO|]ook̷[|o1̏hpa񉣴".ÖGw )Y!œ駫HPǂG"¥Tl)B֙T@ α_RX_u_B yԵMҾk&힜zY[{nn^]-JpѾ3=IEUBg9Iio ,֜I-$gJF,ʋ& $呠Q‘$*fEh _4|р/EhQw||р/Eh gI!)E>Eh _4|uHKiZ3і8hd:}x0p?xlpdO q8u֚[>H|-o[v 45G?`Ώ|o;$vh@ ]EN!Bm0żh1$0Ag;s3 P( s&CZNlC1DɸN0c@ε>kC1FPOD]0IW`*jTaVweq{Jw[ޢr !;=Pe7+}6:1}p C)myOAQWuC>-])DY ט8nz[FƖ^f9-Quwl%B;s^$#GZn:$$SHE&1<>óy;)^5}x>z›Ft|ĩ:bpsuvETDv5ދ~ZeYgF]XWd.P@D >{[j^a&|nh|B˭Yl5I\d44 ]wr8rgȔ9}=MHi]{ `oS߳D/.P=k7HFS0 5dӸ?3/.?>Fܬ=(j)1kZwCtOt'l$e4{A=xyeXmݕom]{λ)y58vݚ˺Z$*/17JPaƙ++V(/ 8yVBBWֈJr.՝йH`x[M1WB 8;JyiZmfM90ryeઃkf%,- ZQ'Hl|l7qo7%8-{L={L)OdZ5Y7.LvNtKmZK p q%A~dPjIf4r]/CW:CW ҕ]`N/FLBW=tu:tei 5rQrQ|A/tS)YO'WGS@&1\|tG8bAZK rZgIQ\'WU4TBGV?6]{cۡ8tY,(ZЕtMi+l3tp=N[;}it$tňܥ 'X]BߟN8!%O]!\Lth;]!J#<] ] JdZV9CWW:CW>XOW'HWrCt% L;CWםvDH QjJQ݇V%s'BBBԞN4#dkNtpq-}+D)+W.m`3tpv}2(/d8ELiRt;e8Sh a}+DN8!tDݿ2NL~X<'fgNyY2; h,/TH$Va,$9UBbaAoKzTrͫh7 hlH%&45kR,re 1ڒ8N҄DYNĊ eCAa)cJr0Zvۡ={*[)g2 s ]!\ ]!ZiNWRiOW'HWreIk<֚6 Ċ4i(Ȅ6b"$"!&a!t.'u4 YU6w-ۖ\)GYZRi/+: p>@.Wuh";XB*JR+΅Ғ ]iD=kz54^Ƣ4RM$^ht1X^I+Vv|Ryج8"H[ qSؤ -D_fh87t;qi{.A;V;11bѾboojPIB1N%1?FVZB=͵{MZp}MaMGhZ=i{6%vpE ծ GQZ;ڢcVmf{ֺzzS z(ϡf) L[*3#tրxޢOZkT167:g7. XIeW 7>:K V/M>Ϯ)n<*.-u6Fէju~/ο;-_R{wK+vdv2 ;&jUV.=+GdIΑCgȑ >k;.JR'JaqVZnsw08u(AV:s"\h- g4'(aqikWi ]Z#{JKpvWky"NWrOWHWFr"]+L%qO%NTDj SB+K)Bճ+Pj@U űZ5ǎZH]ڡd=jW-JyUT1كZv;tp3th}+Dɸ+ASvv5+DɄ+,U!B(g J ]!ZNWo-<]= ] K{WX]!\)]+DL Q=]OWOBWRs#Ct!]!\]+Dl QrP#\ڻ w r ]ZzOWR[OWφtCV1һdIG׫h 9KA?)I?\?M@h1f{vyy\7|+yL'tX[R Jv踃%6E(KC#EFDwʛ1췛l6*hʞE|>f矢yry>(WٴP&aN 8VC1.RmriqgwƇ⍴ɳ46&e%DVLqDf B1s)I܄rP֓RXVO&^%aao~YV$_CL/7z(5nsע\1JޝVudn l^O .4_/(CcN<dV\HioEW&AiC OI T.mًwg+]W6ge~ͼ^A͹)چhL|6z Xo^;B*`C{SB%N)p\ޟ]ƅɷN~ɭT/<:ڷNjCBŃutzEWgQݔ=ֺ ^ &kM@7}zn&j3jښЯzWvķ#[;s;sA@J*𭹏LTX[;@_+%h Q*0ַN^RJdf|E?~WڽĔ35Z}+Um4+Q^.B?{Wܸ B/;3h}tDl爘Ng\S @R[RT@AT~MQ#*3j8)'S>jQjK&{gḩ0 ƨHڡ,ax x,`Kde0h"`GC[`Gm6C ;YԊM[XVmHjþ@9 R@7F*DPm14:vɳ:O;"(Q*V(SP Y&C1muL >[ ^VK88 - ,*@vT }BQAz-A[_ݡ t[b存:8ԵCIvb"JmO[Ĩ[mۍCZ+`@ vCW7_p2a!s89T%h-.Znzpm7L^m,Cē& p&9e0 ")w0J`sFb&3wU7%0T%&Ѽؒe%,;ƤI7Ki(\cʡcclY6hciDzEyXA(ieP~&QӋjwPy\暁s7_ʺ~g(mZvUc&I/(?/5/&/MZE:,٤䯣ݤZdZdU7Y.atމn.bnvVV5b[\>(555ZZ.+@}XVDREys׋yW}s;vʖ:Hۂ}eCA"УLL}}zۋu bU^$WUVd6[ܛ*N+ LwGnDu=4׳ӢC.z|2bJZgtڐEq/xZYRjY,fEu$7AO+FДu#9~+f_2PK+9ZCHdzV>Vp7#~q7ذCw.&;B~S=6WMЃt&7+ ֍c-<~e S2 ;ü,44r v_*+7u<]ֽQu=J ڝ"(x{'/Ń^{].fOTfX]R0ٗ窴M/49{4cAEB1˟Elt+z3hPt`T֗\me,9seչS0u%bsc(<#9 .p_-ߘޮ6 C:-1 '<13{mS}w󑟡)y;yna9J.f-`a!ӌ3J!XYln,[Z9 >\@ Cfr8kcN+Q&dheLiTcV*D\5Kr3A3J{}M0jt(9$'%u$-'й*O뾅أ>;VC 3`nyzOP t`CM%-v1( /w`M~`90Ǩ1pօɭ.&mDlqYDp|b\u9_2$-|4淭d7>sWNx7xs_bҎKAhOa٭ym7>TaX1ʹEoWD /לB^}o-ݴH>ӭAވd# t4]}B ;C { z|=z;@>Gz$<>p[@'¥P Dw@S"}Kp}<Ÿ#=zr ;KaoZnV\$S멏~*H#\B'k pC` MO9Lzf#C9p W`!qϾKu*ᙶ`>,ɶAnL/o[PjQh$IBqhQ&dSk;g.XLH0y#``ӑo9@@Az3fOxo#shԞ!>^K;N~Ȱ Uլmi'Swk7d oIJuζvT+޶Cn GrA`=tԷi< 2[;c)J>Gr 7/y4! +vCB #?~x҇l 3Ib 7:ϐrz, a[dCeLJ@yzRaԇ _[d8HْMpoN|=7<2Hxp$Guj]p6fׅ{I1{KOklqճF}vԪLz6UDrpĞ;by m>rk_i}6P9ՂzݢEK43(%7/i{^wFEJ=KϦeԠ{]9v*}T ͂yYN'eCޭ|0dVewF%u~<Qηo7ta.܍<{[44w$=CdTNmI=wg JU\ZS<2a<%czz'~6jwZ4g-ѱkkleռ݃856a|HEs5@iQzUVNojCEO=Ц]&Qqo3 9.͏YC 冴Xzj .m2KsB|i Bcoڣ]@"%;G^mELsZOl ҨPp:py؈B gN̗x fЍ+%Cೀ*\!9aBK]Jsw 7:^0A|,BhSp<++{yDNot#<ڐI幸!Q .‚loBSSEx- a*8DPONsq!@ DUC8k!k7Śgƿ\TW- .2zJ1а WРe<rDL! :?XIDnI"xof<93FH9El x l6aO6L4 n**^sMnȀV#fIFcev8e8RBT*Џ[vdme|;<K"ȩ8oyGDJսj`x!IKjL)v*.Z s\G/.LmQET$"YAHr%#@3+`RPfi !DWDHimb.&mb.&}EX<.PQOYI#E & c`GH,?8_uQOiY ai=V.'F]Nxɞ$h"ϓ,/dD qE")( yk"R*quAFA)i a_Clz ͅ!`C8a)$ C8?`h@wJ)O}9_/{'wF#~4?/滤N7,!RD҄i?=Uq a|;0M9A Ҍ%*Ndj^Y @%L,Gi T $'P,2ZduӅ`뮳LF)adi{nZ}ɗ5n#۞0`dHR5N$O)(XO @Jǐ9^bꝃDuZԚB;LgjQ NNe6/u&i'^";kf'Z"ݢ⪡)}E7IY/NL/..RO\Mvq6?zeF]"'XC!0͸D%'m#IW %~ht>3X4v/0>YU"i z~^k*msh߯#"7Ef 'P d$Ee|Te4ˌ"#h$"Lpe{*2"=hՖy-ZA;2 #"5?z9j O|u~\A,[2UJ麲z*e+6FJjZ#%88;8g)z,h@XrFS*& vb L:UVo~o e[/tkwwM.( ߛsfQqMotmG8geԂ<" !%&% TDy& q~$4IbkJ "%9ٲDe\`.zu% -?|ZAT6XP*Gd "g$B& iB+m- ᓰRL%פ(1[KEP4ϐ#dstBdL!(}`5$Fˬ p% IݔQ'MDҚUs L^Ec `5I+dIIvcZ-o9bܖcle$1ebhrm7$g݅%O':; ,VW1 eKNQ %_xoQ )L9ZtA16{-e҃@Bb!k!1J( x!!ez&(x0 ʉ"7 $K PI, ١/XJV z /jq05^);$TQ)ɤb^g]փ1cN@˹d YfSJ&J}2( S*ax $x;˄6=]˚_ :9ahW7eȬ׃[ܰf&i5ikN {ݩ6[41|A9 @4LoVm, ,)VHG&uRقÙ2 1ԛuǴ/{=<"W/Fu]Zsj>->LF&OݲS5m)⠕*Zdn5ul %adR?ZeEԂZZ\=e.XJ).}E֜ZdUg-r[ W&OdJslGzqŷ#Nן~Ϥ;DÙXy{;CW\؎S0\SE}ʰr{8>ӏ /㼏Ƥ̤V KYeO!G1ٰ> VB}Y_U &w 8~.7l WF =Yg%&lZ3U0Ѣt-Y)?4=B/-p򾅏T2Uh%"ʒ^rIiߑ]}BUW*W٫-x ?Z_vvsyIrDCfUcss;̿7 VFe e[ ʖ#Unj?.LwyvsPn^}뢈\|f5H=|x@Z̾FӂE)d )$Xf{h"5зo:>R VsJk]eI/sb|l  =2OBm_4 Fgow `;5TQyz?V!#j+F{}nOd4; ~)O?d+v2`+(%7CP/ʃC&Bq?p~Р1VGkiz4UjJ Ċ+r}K[xE h=`e[nnwX+Smc`MG0JGEI{vs3\aT ޹&ޔ!l~ ̔~y 3@n41:CO8SXdIf&H{k@owbcYcj덬}tP h-R0l>VxFT)$PJ=_dIC9Wә'zBLhuv8?!w2Pa.nofa 6Үo^%^, ?XWui)-3i~BܕO_?$(gƱoZI}n/]勆(U+W\2-35ė\XAWmPA)[7L$6|UT8*`ԙ |BjC jT 6cw̘FhVWo2JOT9/Z)ʍߔ/uT1ï >NgCs8*ksU'bhFf id!\+fL6[_H ͹g2yjl#pˠ _柈aqp/oQ6k6ඦpah@1zsP5vFB- 㶥#T!iFo?x +VZW#l){;†{wUŦyO]?|Rm|i|Vk 6FCW}2 d7A FnTiR򒍳T=;DlR϶GfS;L8CLpx.@4Dy 8}|Gμ~š.ȡSLDzV=DO !`Lys<ȫOel ma Q:ټ 3l:=i\ k|"\|wYA"15T"q @%*ϣŬf;hc{]޻{oyxلt=c#S60F'`|Wz%Uydhj"gH`MR)m.7[a -uYz[l80Z\^i^ݼuOLb47dVd ^ĠQ|{#*;Ę!ŁsƊ 5fa":UV]ґ㰿&@TE# |2m&qj)=-uykGZޮ{Sbc ^8 Ӡ o<όViw$0ܩȱM}C)vUw)**豩GnyY2ްra֝w6Lz9p]Y¿x'a)Oy\%1ƙMpYuӭv=9ǻ gu~)ht_Ӳ>YGŐW?[ /qYƼ;Y{  wvӿR&ےM 4ӎ!bڼQ$0Zk<`!tUuKQPwO̾nlcV[?ڸ9N?G<)_B MD { s{kNQ%?OŲݥPpo5G"COK|t1-S_=F`ι\fGXbW+cv{/:7J^fu:UޖKx )+QĄr Pk{ۓg,M#1Ʒ&Ljc%p%zrΔ}D)4dT]oTD`81t?k(EcH>ߩ9 snq2QQRn>H/-O0(*Q]EvLrPٖ 﫲ok`mgvBb׈*q@pc`Ƥ~tk@RJ(3/ArR|mLiDhwP*S=γNc-DjfKaIpҳ o=DhyI5Z QKm3$z$&gkYI] ?#UvyJKI9fg,dhf#&tWLeDTVRƹN?t<&2# ĺ'=}$j3^?&y(/4Wݖ 2gF:Sd@k(\S1x05^5H!}Ӑȁ"X%u}ןz9Ψ ?!$/?.af4*$G[Eˆu]+ u/ pw Zaq.q1VF#G/^S)(Yi`""ȇ;)U2Slr~zvɯ!#yAU4{.L6 \t!r+k8,ۚuĒ y}dH U\*wnoq aqVNö5,ko54z?#(qVuMy챷|2%4zZgi7IQGbC*#!*B腪(1P4d~*뻏Hu^HpSl Ԍq<]+(QA4ar DfF7>]g hq At hSAKD5c΢EFwB:s<LJC :ۓ܅;zK_*ShRJ`$8xӥ<&;n099\Lhz5/c(dT`%<\?v= + JxYR?if(& S[[m17XٽtERm #GNN;W`n/ gȽ? ;q۬ج&v/ %C(U8 a8,2!3OUqSWޭc1~L8;l4}4y +NNcf=Iq0t']%RnS;(mEk*[SJ.ܺ,J#$QLjy>[z,҈dzP'3El 3ǠsdȜJɥ!9ZT{'6kՈOA%ԙ|Oh,''sb^^aWeJu͖bS P  c\Y*`P%021ړcѻy"V%W'@e1r1{O"k~/e렵S}bOg Y;o|zgc/[)ztCeѲ\=Z8EU^pGN% UkN1Y[ Lir&\ Ͽ\1LkDE<xd.'uU=ܐ:i0` =r&ڌffe$U :oHڌ"$kSu%Yx)Nⱇ C{%U2.{ ƾ]: ӊbx8qy9:a氨]ލf߮dFA5?:[ /uSAw#JRRRuջiݲ (5!gUkW4j)^^~׵5"r \Q%ïςt_og׈.Ø U1{}^[oS}C\9` 3S4{}JuLd{U4 gtХ-|55iᦇKdVOהּ<#Yi`n!q|(gV ܸ/JR@?)uܟ3PxIo႑!ܒ [\kݳDgy }G;D`rLFz6H%Mn͊jbĿUI*矐o)rRsT; ۿyLYmOJ60Vݚޥ8Һ?A?߂0[B1xBȪ}Jb:^Ý4>k4}BȒl2XlpdZ>v<߸F9~ȒH|fo}q85BDFCY !IT'a3trfbi RV2眺>YɦSk у*Ƴb}A\"6`h{/lt1NY|IyhZ@VZerkHe@#$8{&>&mj؄Orl FGU*2`'d^ :T ޻=_%|z #Ff<X_ AT/RMWִ!QćIUi8B*_7FΛM[TvKJmBp>SB*P6aRȯyX,T>VRUB( rچj2avYx?,XH ;K1geUOT"Eh^h Rv͉gx]DmC=/lYc䜩p(b EY[Œ"nF2 r*1Բ턣1zqպ☬BtH:0F9^ њړ YaKT{~ӑr>oA%uk*T†Tsk7_XƼL,K. l Ͱ!4449)v\PR ddu^ ,dj? GiWA=F ]d<kyZSBfb5[U}ҲƘPT>^ʩ85A~3bV^AY.sĔ)\0JPJ~S+ݹQQ=౉#~NJ!g@vc{u}mydGFKI3 gP:1&-br'aR ![-.уؔ; ==HTZ M> 6O;] H&Nc$U$$`wѠԲ04S +ϱ6"t ]"Rũb:Jt؁&T:w*g/_n׼A=ɳZ_.P]>XG?GѨ64*^_O :N4ن_% 4Pw0T^v5%蚫&E ka݄4▭!Y"1'gFC{$)ęQ;uT9W*elEY+ uwwM*wAv_t(\zCXY m[9 sK_cnKX=ZqY*[]  FN_:X/Hq*)8wITmhTnf?G05hTX*Ӹ;7ݩn T?b]5~o2qj`\s>sϠ6~BXeӬ$ T?f4yor j9 ȮQ<,z`;I=L:Lz,{= Lbsgߣm媁ѧttģ23z+]_ߧҒqČ 5PY!CYbV"Uym&CGҗe*}`ƻhGCiCD.F½ϋ .d T?%î\FiFF]IJ@1vF_O7Ea5RX4?NȥSpNikCk=<ϧinւnE@Mlapc*hl?sUhyGAr56i,wl ]T&โZT~( 9lx\K a̼4>yӰP~} -a!dDtjtx+efL u+o{}M 5Wu}8Dglj?7Qth)Xo^6^6/}x¸~I Y#\^GE/@/D qqu l@2fVF+zE;Bŵ:'TƵ]Mw5V7_4r5WgrT^3ٻ`(Q4aLB }ƽ3C.4t9,e>1fVppjaK!QWtS(E,lEXG\F/JT2YHP+0F|x+Fܯ {q[.qO}Н;+(0^F>om7< 'ob7Jb|~B eY_)bLoÝ*\'[#Lds¥{ [Taߡ 3=XKsVE^yñ?idQz?~@R Ee"Y-0ԂIَ $Lu:zK}>I ޥk(25y&( .Ì\PIU)2L2H)rZ:tm߹ wykieA#;C($z7ވ@Rltv͖Hw$O-;>bBr163*"r;_3H7nTj [PEzl1PE(JkEEID_Z5*(bQT%2kVt`w*I| E"9$ODd(y? ~-^5)crqj-6ۊR;dEؕXfMlE}\ER55$H_]I1Wi,ǣT\oPI2zu7xiD*ޫ@>bp;q5! Fc/H޿ue#75d̓U< j.ߓ<Ÿ|I̙(XF+yE~ HJyiՌXu%?ʹX4RW50ͭFX*o՜k` E\*H&QmIZ{4 Oa68OTHm~]]}tvR}n93^8cٍy >ކ]FQ-a|@(GSpלjno5TJ g-5Sg(5<{z N/UT8Y^iN0&X_\/`(8;wh+ (r5f$k`\|؛2QU} +J>fF)ͬSm0EP`'`N:qJmSphь@Iz#rLr^Yl;m׵`D߸n-~錏[l[q]jIa CC 5O wM wMnb2Vj`E;tPb߄j̜M[S S&UE^oCwyMymKwچڜt4q*mR_4䖎T !aI$|Pc p7gXXUj#37F;ؐ*Ky= 7'-ӘKbJal;0tq;gn@!-sw'"5 G*-)BPGyNjZ_[*U\[ŞĨ7M;9h)+'iֱ|Lޙ&ҨF<䲧)1H<}.(sv˪~}9Z 'W]"WUbt#iI1XPG/Eٗ&5X9&%"ͦyRA>$ Bm_*9b[=~TŎTRtjᄥ; Km}LK"Y(:g[2>Z8$͓z[h1xs>/A z`ӷ؍&.~/._x7I^$fAvI h(u.v;p'wʦJ,+Ǔ#!I#VWGL˻6e&aumUIKIj\aүhW|r^ZhYOG1||Ȧy>ɛLso(>`^|5;_?|mQNhJtJ h|;of> d~[cvc3t?;q7@FGధY28\)k\B Jpxҧ*'QqVRSZtpu۷!%BmwN9b=w{[V`hW1>Ms߸|jV`P} ( C+8ST6~ùC|%_!h3ߊ 3^U6 ND8};)q He+N C#}ЇJ9"jd>Ċw6´8>ay'8O u ~De|>+`!,0H<8xTpC\u)i t5C7ȇ& /jWQM lb.e@\h[ʐ~hAqe)_pbE/{i~X0Z=@AD37 m_~{G D*1~^"/ `7,=p}ʲiv eK{BS\ QeQ7JXIP)en՚Ѭ]Qs}SMj%j wUwp}9M=6i}}WM TՎ*5k;ra98iؙs9N3,Fؘ|Ӓ"U%l(*"؞ (3F ɒ=×.ǶyhֺLeJØJI, M#ktpk{ؚL>;ɯ!\342{FpIDM%XBՈ`WvTz$cȔ% S~ozh B5 ')cj/j0#gt4I)aTO0ŋK{ alz|>\4YW2D QFݩibՓ7hh*"Fjšni)ź̞#]؂2)U5srlq hPNIwwsNYFP`R>lŢuuiʪSSlOt[x^I;+ |j3Tʦn\효fG9ズ592BQT/snԥ9` a!$?`nzT78Ns#+?:K ]X!&6QAq2ϔntp엵sI0Ѵ58A\[}^~}yh24F2"> (}%#Nq XNEt{:![2ΉcbMߞ̉%TC'$PLRc8lu`'P Tr[i :~[PQQj릱N׷cLT@tHXD,a&p`50*oc9.B0$/O6ل8ƻjA]s-kƙdY!gks6ܾ-1(q:X8)->.F,2^&#&ѿO3?-2~,?o_[_2`H^i{{۸)~CEF!Vgȿbe#Ƽ %>YriWf>Mo>s-':P)_@rx@>)E)IB`ia$YTX!*1Ҝ+I&vݡ-38^:ZdEZ>A<bERK@>_\'KV@_G7[?K]4kg-bʊӑX*F;Dž~Tqh(zm] B.JQoᗗWfkҁ]&zg/DYS%>=M' Ai}:>%sA_&On/Q `ga|'qDPhh^FYo|E6eTN3{T/0 %5X'H۪cg1Mx4Ͻi-p&(_.:?])$Xw|h!/wiNXZAF綳,~=LK)/Eٻ6Uz?d0XXsPXdj=]ǃVdH _ ϠyӌKrُ>-[d0V۔'^[diZBUqi-~lcyqcM|U>KM.$'VtMiUQ}휹yp~)r-,A9 :ϩSQ(W[~O 8 B6NKVZd˩W 氏{={,x/dJkex_bZP ^4¦X@#!Vɨ7۸Z(Ь?>"&ePD1'`M(B8K*JBݺqp)yt Tho |gUۍL5ӄxc[}vj[AV;heNreF1>HdlHbcSsv؞kE~7>4s%L6)VI=aT$i<_XꔟWA<!h@Quwkr2?zk܍4kzŪ]w/Jγ9fzIkcRG.0~oTiftmUG&(e=üu[b545m]fҩzM| b#5~t V]oCX࿯q"JyzBbfֺߵZE,h'1wO9eؑ$rL-RL=tٸ{1/1/TN+CxU LNPum`+V- ) "tt9I;t5P;Yir irY$5&q ,p 2yr[5 l oe*!xk %I+asC]hV!RB6P ;XQb72<`v/g-ya!o>fN ٭ziTdM_NZQf|@51ycH0PTo3A;h~ڨ`L?MuxMn২Iɏ3(/d# ؒPt{_yHFA‸Vو 0Szsc塽fS!tpJ~R@jۧ8J~ư'(VQe[K9ˆAl u91}HUjvX̿liJ;:Ӵ2޶7Fn?"`prw׃Ko Ki4؏F xBbzCj@ P!Bm?Ҙuu4g܀͠i\n%{2iOrB+m-LF)yq꼡%~N^!0-=XYvu.87.|];hńkӷ˷Em.^79*U ` nQb< cc-pY*=K4_Z="%8d1(@[Ϭr39^m p.D#y}(+UdM{UGHߖ io>!Ԙ Rn:Q|P")9GkR̫>?ӝW1B}{KZ%^;h˥o;f 4_ֶDtS (ޟ;Q GxCQ  'J_lx@s0?ɠ%  ;jcVKA)jKJk, 8'WsN~F[ 1ٲ<VYg`-a l)wRPHe;iZeC!;PBfj56m\F:f:kg<}pDQ3xbAcl𬒙+f2ޜ58;5~bx? *՘b$"},q?zFb+Hӧ9㫊TĺMEz?\ݒ̬"EX;`j8.m T"4~8Drڧuc$IfQLv)`)"*x ')"|9PF08j"`]lH92r̠fp~l%1htఒ#V<%KRBhB0):r?+l_N<;* sIšMN:brpڈorq`ܙyy4'b tKǾͻ7,˒nz>ݢුxsy9y& ߼Yv8%r9$cFt_ϪKZtWLݹ=pvRgَbN\J2mg'#NF tN% yYwpUY{[b 'ʜǕ;dB $5Qx!EMU<6X],A'$O1J#Ds;] g9ws9pMY۞UHJWR-kE#Qv棔لꩻƔh.BZhҳ;\[Ia͋eώ[A|r{Fw" B蔠aRi,,ޢF-TNϣB_5 Ѿ]`)X!hX}+*^Ft?~MZtpCu`Y D5 'dMSkg8MHR@x+ro TYsc6R!4^4GcA[Smn `;F>5{Q8uV`EvKҸS=ŇS:<[,ޛ/n吜V:FbZ_/A^ 2  2۠ DV4Pa6^D;Q(g# ´ JY[nD(ԵPk͂Pɞ6xJrzޠ3"ܜla?F^O>lGi\_xxu?v d''s YNt?C [ 8ڐ8)OXF㹼j9P+{bܕ rqiPR1G'mP$|ҊJOhЕ\)6Oy]BgwgHD'(ĭQb@ ]%+Qk2Q>Lr*!`rQBe[#r$*aI̎'h0/p4wJEWd&yR#Ci]4^Sza.DQ,YDF=1B"? ;_ؼw9H|;霖ЬǛ6M'J$C$+m#I4fS+0vc0`l#+K6E"k0}#IK*IT2"3D/"L&hBf5jP $$:gq^?Biw[/Di?x+bt,>՟|418VrSxc|"uȎ5nXHh RJ4Ux߯l}䁥ڱM>kʶkx;.>\r+(b8V=(^l?]m,bmҖe Ļ=&jr"'|86620N1H\84WtBŭ4ZnAt+&&XE;s: 1HY=$ m.>;8Hؐ4Dϧs܇hf__4๨KHK|@DP; 喔_ӎ,Ԧ2l;΁yFnn~B3ݶku ܗ|.I$#)CH#.&h dw_& EI!u5Yi'Wi,9 [`$~E gSFֳ% 7V}i*6]4cesvp; >yg ;{'؀2r_H#и90X;;on9L.V[`+)t)?w$Gld6Og}TfFk>4y#FH["1O:Cljd;8h 6>\օUW0/^#_Aʙj?(銯7)tgߟ5 Y%ԠlђK-lZ玤;!.o1[I?n~ѿT_L% s |)g '/rl;F/o#|ͨ^&]jꇲqVځ3'?C9Í<7w~3rk{L)1WL8;{wyYߌHC[1_Vc&1i4M%.~ӟdR<6u)44e3w̞;J zyH`>ἃKO8ơ9+J[nղwfŸ VW{u(+2^9NnqDmarTўXm|@78ay& D“A38U0Fˤ,^c}9Dml[8bIIxBbHOBWDQ+!wp*f';ָ Lč/CBNķ&k"uqަzO537{"lzLMY#Oz[t_&/߄F፿V_5o8_xsv e`_]spd /=?,@V3/[s g11+m 21Bh"JAEN6gg\ޟMȗiPK`ቓ]˴M~5F5yVFCO6܊-f2dL-:9V $JIW)~ݗtL :/ohd>=PvWni8YJ%428b4BC#kl6qCd88N>C\YwdɇpA+\ZF =ЇKԩߚۡR|y0=0K/|ps׃[}2PUm+cJuV+mZUF0#l,)]a.zvq` uG!ҠձpsyY-M7ti!iE1ZewҢN#OMR՝e$42]4j-#OZH0\qlw` %נ94kXvKGfv.O_~ecZd!/$V z2J_}e[**++{fT{@ȵL˾2lKV +,jO@QZƀXR1[ƀqE?Uƀ1`QǢ}޶71L_pIX")e4}FahnfmQz{4dס˾`B%Kzs&X[R yo[J>lPt H{mرvv S)Xџڱ.!fe^Pg*z[3ؽ2wƠ3"ǑZs9 ʮR1a ?h襴87hgɃdWkGa)iY`2wtk^QZcD=qBDyEOIy,+:)aNl\u]SmŒV.8Ηym>^~_~h4^]]-uko_8U|ԇ=~+ٕ6 l:f9F)%[{0 "]I_d̻EGU\S4V.\#A53rЁgij)ks{hŴ cz֟F;8tg;a3Jqջ* X t̛k"$LgR%~kJjp#U(r3%Y)J*ބ]tnӻnDgZZ)Le w_Wr_>rG_ 㚸I8lhV@y@{JvqC50C2AҠɇ>;5YhEU6UZMPFOrl;8&ڠaѷ$0+{w NG6޻8'~-UIu29kymL2RF֥v_ǔ}kʨc.׆Ij8ؒQ]2w2$U9HtrX0Fx& *^rZ"U3)A[KsVVf6dxᓦZj?yZ|3#0(aSh(*Zh"l#ʰ51jebN%,AU@g ^ )eZoyj,-Ю@{DNooΌDd-0a2={/)bd lKXNGq[lblo"p5b'l 4FjTP$V$M*$h7.(‚{oF3{o5(+[)o}wGoBh9Vي^9">2-QGu۷ip_FKaP-3T'"T> +c &xㄫd L "TqhԿ!ZJ Dv.is`iesЪcHHY $S=ԤJP<59^r4_,'!4!H:t% OxXrdM6GR,+ᨿ!s֒Bʋ6E3x}^F&+Ԗ$cFh]>s3jٝlݜmUOVFxQO2Cx9&&zjR& IrcHP4iNvYLp 7 = _X؀8.m6ePLeuQyipԂ`XmD2? `LϨ~T )*ɓV#JW\+JݗLD45 *V ^RMoGӺ_ph23 y Z#GcS(WQm2Qh+Z&0YJ> MϡnjwEZӁ`# gř aH| (O:kh:kZGDi[x%P81'ESlE-ZF :-NFP<`C\q%AJhlpSJ0#6`iɝqUBIk5NiMKDEQD4~<1 * 1f;f`hK]ל95Ҝ۞eKc-Fۤ\C H=$77Cġ7Bt:L:MRZ4*v8(ط`WBO-weZ5j`Tt(e~L9UYތS[ʘ*}N 0ˬ:{InnL=㕮?5yf9VOvWhGt@W`D՝$7ҳ4-O+!Ay#(^ ^F8ڹaD䙧>)OMX s29g7.2jҞN zv)n_~O_ߢ8_?%s?B 7v@`f}y'܁E7y1zuG\To%@S2m^qfN|n&o.- -m]}v}͛)36(gi8#g6 ;=rLhxn}o& biYSn?>ŏzGNrJ͎/4) ,4tǣ+#4wC^H̼/iٱeXȜ8rJ *?zptD#,4kR"ex_>I[zb%;"Y`e-8N|^m!su=z6І N >3}#rĪz_^-/D*MGi 99HFRta΢C@:z^A(HHQeԮcf]v|5f!~bof#\XSZ/S9RR8ir ֬>1 KZ'0TT*7I@w3A ڦ}Cwր;:;6 7m`(sYv`^AMXEqm5!d&џ:b!5MRI-⹷Pee:2&qNfg\!h͕XEGt%NJ3|o;Wz Ԟ]9ZzaɅgMOk<>Vˠ)gJ3#T=mٽ=W];cmZ?9Lѿ"uQ]߾?c{Rwo=ޫӚޮn^^;#f–-\9Q= xb.q CBRY1>,&X}A*(l{KFN;j`qSʖGhTqt>< 6Y8A p„GW&^XLgmSuB Dպ^,U)b*V9-cp R[fA' Wcsך7+&X|42& y(VD/׽ŵUjlɠ$OdR‚̛G$LI8B_ul>x-#lJzs$ ( OC PBǔbe7!]Al9;r7["ע(kI'Z\PiO;,C H!BF{ziB(>PjcސxXoު@äR[l(~x%3*KK(Y5իaYXU%N]]\<'o+C1:^QGj! !6&7PǢHT b**5rj^RV]U Fzwַb#WԿ PlkI â|!H F%GH 2Z QUfNΦ$!s)Z>[6(X*@VP PDA#rlUJ, tt3%Q40B^zi'dL1)[ps-yXes̪n%'dfop̩`e,}*Gt#]]Y]`VY-5jK6ĊbI]kJ%A?"䘛u+lwz +Ζ&-Qe\c*AFWeIj\}$6cA |w큢\ 8c4΃}]6Tq{GQSAUj.ɮVIܠ쐆䔛ji@t{YkRmekJsZѡEbQL|DW:+Oڼ ąerƼ}ʜ2LoÇKr 3 uUk]qSy6\ N"+nF_غR 10rE[UJ_^ш&oo5L[;L>F0f,PJP0deԄQ*ĆjCx&%s.eZpJr$%@UQ'Ad,(ԛ땀$BbДu\uTA) Oq W˞F7b4@ ĩMXq>*I(KdK8El–nS4_ -?|d(/zaPu6g#)-zB>%eJݒjgC"wg׸v[05vy^$ +p7Ef%ĬBՔDڥD9g'1:%gds6T\J/KԜ~Z6B>Χl&=Ē~wH]ޮeON% v̷1㐪Zֺ߲G]VuCRΩG(NY#Cܼ+׉jԪ Ҽ^blb{Chyv L'Uu#{9`OP.wINjc JţőeX {ܝC/#}s,sլ=x f + D9xhT?c;};~#HUgj Y6g:#hЌTs hy6O=$İs @uC~ތXdʅ%о; ā>ʱ%|OWTG7V{:Ad>L'g^|Хc DI/^'V︧}m:# yg=cx ɌX:<7_rN^ˆ!g+31}SM!ء̃?N|/瀫gBxwx[&lV/FuϑRi[MRBC}k"!"etr<yr8(X ~2H&x{;uOmԳRk=PQy"Jٯ:FwԪDe&Pk&6H!c6Q`ceUz})S5?f]nxyqZhʜ+kA|=uM-bзjQyB] 5YtŗԨ8&5* )A3}`c,&~oHb7 Lw>1O5%U_doGݘ*zrJ=R9iDJLPdxzUne{gQWktK2{*gizӄm|?I7bOu/ ./~r.~٩`$_rqeETHiJÕ^I"rf+ncز KC y'EM%i4ګOږݯ3`T)hQ")g%wuAWagϒa0oɢgfmVMb`gQ<28N_lkJ'TSPO9,CW~QY̏Ku (|S2!ل?9XN03ư kA8CW%PiJu#)QJ>47_h"< M y$hŎ,BӴ} =w)ͯ/\Hoj?׫(zk7늱l- $zIdb*hc,9"Hcd.IN <5zk1ܡ{v(+:zUƣtYky>8#A,D0*ivVQG\Pj8CP&Te+p: †N/|wLo>"jÌDP˓/V Wf J-9bClKOA-=df,3E~&9> 9[mE=ˌ$&PIh#7%4BD+ݘ"3)O'/y_ۅn2̪<;;;ۓUQ]ޢ%[G _括"52<ރovvD"a(c0:я Je(Nܐd$17䘜͔G0ep/v2(TZڝߠXxu` Lo~opw LWDޝ(Igt"$b jϢԆӇ{wiڃݨO7"GkbFt J`iJc(%lb:&UNN;A#z]KN}r`C:y@/%P Ƣ+C]2h*Q1&}T-7|Wl3# Bϡ3-fzk.G(<&+RYg![GMR"a04UΦ3$f,2m( 5-gS!%)F уt(]sGolߗ;aW%Ua^1%48:RL-E=3J>6m)*8˃ŜULE1',sF_b'IJp3.j]du?*En?^[b2eoer .'1Z@ ⯫q2XJ67/g!Iw"j\O;T@(8x1 Z& ftl@E)I)Z;Bk[ΑmG#LP2BHXVh1K}2STFzcmь Ge2>wLI)`5 .Xʡt9|68쮊>:4ZEHHip.hx U$)s-(cQ{$-(e=iU,d>AEe%&2KL plw[d>Q(`V&I4.DLtaQC9׋$U$?TZNe%Hz8Ɂƕl2b`DE/j.΢p9T*{%hk;ڇ+!ۆӲ,WR_lғ3nlSa@+%)ѓAfa3PnnV"̖P$MJ.QRI*(ɣOP9hY (rIE !QV̤hNc%A!'y;Jr\S;X ꈂ"UӯwNeĉ+Gg'H'7'/V Q)E.0z _M' ۫[m765V?Af"1,&LCm0Tr]ORX*/ %k^# 5h@#${9Q=\T)u'/j#wVJ7)io rZ`z}[\} "N_˱UOߜ6[|~QhtЎ֜*^K.B~OwZ9$^2@%pA_)-"1#fJrVEY9e>eK &Gë@FQP(N Jx*U Ղ\k %m5Ytw('$$T[/~]^Mzdz}hX)ohͅl=29$*P]%p[xSiGNu/E#RR8"K;#uR,(oL,;m1 ܸɦlE2D"#u8yT/"0{3|~ 8z-WjG4_"21TtXZvuA4X#3HN%z-nhV&x-<'tFͭՎXr^ ^]FNJ9)i䤞F*ڠguLX؟a%K5SmVCLpfGj^F"~`@ftZP\!0< E8\ǔ3bqYLRɅ}@JLԾT@C3.^q**g2s>%Hܼ1)t<5Ԡa£1{nx6^կuZ1SN` \!2 2s]L˩P 0+ uyX,XG掠 h78TC P ' tSCD\ЃE=#1pKC"zCT@c@o*KwosaQLȨ81kd]:Oad`ivK_?anZ@ 2DZ';nienߑ)L~a/ڻ4Ն1^R\?Oj6Jt:u]؆1K*\hiT+)-a;V8DT p8,YTY6<(q(hL'xLFDEFQALgMm"/ ރIDIyA{ /J=)Qo'|cAUQ~?D6u)WJs'ۇ2r ]ki`m\_{udVV52d`Zw ]b;9Ƹ榇nQ~]rR%Imd3eJ a$"\埒\vg;J60劁3+*cI;6δcL_l ;f %G^c|E!1'1`PKT9=q,8ްGˋ˨cQA {3HH6cBWl1 bS  UTH &h ONxP{Cuw(i%L_RPp8GX2tƃƖCG|x/rPY(ud/,Ct/Д O~<艉gw}V`IqCl^zhnήB>}{~W~ҷ7dz;;ϿKodqtFu‌L3?eO A+ib\^tbDvr2+;90qߑXƵY :­u._*/&$O.}'G`GcThꁣ[Q{Mh$[QY$uqu6D~x e`Li[1nik0cDqD͈C~aCi *#c'$ANSb̄ 5nHfPuA=Bp]G 07[aln`S/ʬ8JlyWh }⟛)S8C*Oƶu ˠvV修Vz'i.n7eJqD l %䥁zTPc V۝MY?v?$*ZJG5ljwĩ/@bLid#\kŮ!q#OgZHrb]'`Ub;`3~=%']~,Iz-$@TjU,GZ=8zKsF֜۹ϧsHݼ}ʅ8jb {WOLiދEּiǭo~^UG&Uli^%2Gu+7KGHt6reйshkm7.Rs:IcS?nƼUWni5c9命uUEJ˒gwXM!D!<;sCmӺ]'mZMxoŽoHֽ[ [oeƩ1pI1j(}Rev8^C.+po90-Cxf@7{e$ǒpac)2m>jvErUW5tv:~~ l$*ꌾW6'B~bxcfΦW^=q!kltyީI/y?/Ok /jK?rok{BC>~Fa }Co*L"H fm~#37spvBh،|X0H(P,~fRQc~x˂D#7Jfe/}:`bE['yVl>2ÂMҭf4mtӸqdk~K .&IRihx6MqQtaTi61֙WFX JPB *%E@.*cHBzEPz.RЍ18ʷ\ok|fn¸XLyw."I5 jΫv0Y]h*>S0jؗe2BYlQ xEcC=Ư)xŵN k#7CUgd|=z _Oj_9\B)磏nl[4?}s^Nͪf+̀cyZcfJ~3?#vzgWp=xqfHgF~<T'4'IK8K7ׇ|\Ρb iIquK6/zB?rKt n֘ V(EJ|%]mHٖ3qT-a,:$lT2&1.9u?ς4DDuM0lWJ*xQ%  (}=ƎL0o {UL8 9W C..Z[ڠ\sQ[ӹzp){;fHd4k<MLdR*3zqCJp(7P@0"PkإryW[U+z"g{6ᚱVkhW}5Kckf|MZdWgrV N-49foqB"gNꐵ!_^W6BMISr=iYEowN9}ӎgom@lHr8 zhtf=Df>.fO spÄ$vyp=9鄡s=aΞس=?Tf{;qy~&{|;q^\U!NJ0F=3kSR>rμۧEx|#Zȑs;35E높ϴ5)GQbޑys }K 10^N eJ apS60 a G:1!bRfW *Y))ςs ݩC(v8̏u !ڷl9nQC9TǺ>#5;LJ-qyzteeIPэl(|6Gl@Rvd1 .0MHSAǞβ2P펖,hy;.>c1(dq 5 p $4 d:Y Rgц`O+D-Y[lR*ʌ1p9*dETc$EIp D92:2QY[0Lb7e th]""PdtSQ^d|V2A6^bjNES:Q—>Q*'ɾ_Ic$S rl6ؘo7QrSdh8V;ܑ `L`ۅ B Z%&Rf2- ?ElI W>DJnu$[1qdZj(|~>zUg4F%TyEV`4ֲb0 \zڨ%VĢ@DESX'0$1/{ HmeWT)5ɬѩ]5*5*oDJ/,iFg5LWZc+-=]H3Tr9HNjUPm"6` `(r:V7̺=1ɶ0I'y ٱyc # bl W-oA♺w9!CNlr{YSj\DŸZRvu-L;^=HvAWrMɋ'JM޽91_'XT{YuPgqٟ׊yX\!aS´Ӻ$O-UrףIcf|g#WϽi~09U ёIk~ ÿHMu=Иw}}iq3fU7Ej]y&Kc|;M9xiZ9GƂ%{X~R #]Ʊ(T0{c$9xS&-μiٷc<\iG+2QF#632tC´A 14؁PsM|6~Ko o贆`j]2Q5|~yI8d(Jc!/0\d3kc}u8O[Vpuզ30峟D2[մ>~~jX48 =Vzn-lZgxƀR7c+FيVXk%XiVꖒD>L+eR๾JXq̉UUf7/QdeCP$TUgB21|)$0J0A-R "iDC=B¨x "J) Ө`z ˮ%R-/*bR&g()񂁯VOV0 'wT>-!5!$k:hBxbGt($hj_oU# "K]AUaCV(Ť!;RX$E="1UOOF@@FXc}2!DYV2Lje#$ ,VF!{VkiWdC:S.Cj [:+?"!KG3g~ cή''<" CfZb!avxY:m!vH]@:ͳ S_I2mWp mŲ$K`=#5G=3i:c[&zWd$lH8& 92%[%)K`Uɡf-p!LX+c4б 8LY#ee a[ofc ~s3H6{ZlS dI$Z*s-|VHl(l0250dHnLJ`퀚7%DQɳxuA<|U-jRq]V<{K15VE%ՐMfU^K; tR9$ Ƽπs ?}ۀRM~E%byŢNu+$O?h1Ԓ< h7u{,Y;{O'F{iiI\~aQ=BӤq+V$,jȢc>iĚը3VָKԽK0$BR/=]6hK̉@4R֎|d)l);eCJ(&tlgo8:&:k4|pK[؃xe/> 1=3Ąw>{^'`K_?lh(r-%tRZFRldyz6j=Bk'5|8H!6yN~/a)%X975`hB׬]OB-b)@$gX9Z-^lY~ qrj^NbΒw<1F2Ɏ G[ SEi Tc (-_ͥ]żV t1y晸Y#$(&Ib-ռ,!V )) 8𻃁 RQk]IrA!!HUx1k6&Z Liddixu|4cKf;# ;=^;ܪza*nbM?X]eSju_剩¶YYdcXHipCVp>} < eU'+{WXFmp^+,n lT8`Tk NRYKEn@qdnŽ"y{~)lcEriB}PO6JwGp<=@m/;Kv;ar2Ԙ@:>N19IWFWQM+;3e!O _؋-K s˼V(Ţ%xb:gc*"cpIcl"WKǿoe ,1r\!)s=Xdqw2G_rmkx4Iu`YFǓ'Lah$A{Gs#K,M[)vE^w>b.SnBVHSߜdZu~DC; dK..n^/^W}Fx9}jK(IPT"ڷMdtE@Fu0|iRoZhX ,x{Q>^a0Nq'5/OV9 k]e%$7>:w^-Er,瀜i.o'6'6xW?O4m9_)_}>zs5qz kWLBɧ!?KVAǯV_L*>q:!]6˝o^Orkh*-KBg ޺#3Vy>sr''np)ۚtF&7ԆcnrKGPk)2dDolcv|F;Qk'wT2,e+Î#ɘGd-КatoYзDC1Tw6h̦E=pU_{m@9q׆7ЛzZXYׯśr, ~sR_[C] s}dM:@o>]~8hUP.c"^\ f<Ϲ)cӱ[Xvv J5JՃU<6AAvZ8vwƴ9Ypcn=wAIxJcrr/׹hz{DW84<_2eB};|8^6g7!^:t!oP)G0%ˆ֐?:T3mvğW?hd2pAQ.@GL*lƢU,dU@K  1(!']^SB> !ۦL4M6eTZGD/+-ADxөG<7*Ǐ5US .a[OL&ՀC| *%&ӽh섖'ep" F/ ,w% ,.&%ZI1KDc@LK^^d:MP+bm%)PѲ^ܦ' Ɛڔdzgn bޭm Ν\J GnY>Z٢%nd4%$ϊSe!*G{㋉Z`r5]GldlPW7Y163}UOMo?h0K9tyFz4sIO @(MjoJ63k3Vie`UWUxР f )xEV⵷ N2o>p }샆B06o3485oEQS >6Fv&=)IvkJ![gEI$U%- EE6A22_9zSVQ#6\3f hn$*gÂ΁u.sz[o2V[Rc]G!1xW ᙜy4DNtAZT0Y[$"E)b4EjܞLkSK6Kzvwh~셡TU;@(h8RVn(*cCaBd,g !#SBnhi>ۓՇuߪd'A+n6jvZX m񻎽о!gc͑͘UJ6hvp4J?dzKjwݚvNZz ^ ~bwIX≁ӺJB 5RLrSj$_Z1fW?U89#} 3Zv;Dk)Ųݎ S0$ɹkb`AۛDw Rh'KQղUh\WYW\B|ރ] "D̛lH.ɝ3%q%$iDU hniEa01o*GM8!ڙ9rpUo&/p5 MYF wr]I\`$y*:TMO dpfAQRV`\RjkV|V| pc: ]0eVzFi{õ fU#pFXiViꡲ-s5vڒz@59PnemcSB! 葧˟N3134P/C\CАtV='$ݐMeH-ޗl2kVHq+$!&_ ?Ry} ]hq 0Onv0#,;\Y~b`S~r̻@X*ʧT*U$r-WXUb@70–= cl9r|i%+ ?LF d[W\uw<]*5#WC*TgŸ!@i$!ysF3 U/$xjX\߄jo޺sQ L b Ax$3YSuWj֒MyTtKZXv1:4\B{8]}h^5,]K/+ U[Wyǖ1Z˼y]t;fClǖ6 Ėp][J3Ė- H!(c@9Ix`MRBsi;M0>t{Q'eIktU `ݢ(.}o]zZ[.=pQʘSc*c~.c.=0zb)e!snĨ)hHu.=z[/o'l`.8tZ\Sr5VOqP f`.OMPb`.zt 1Kn-_yZNpC'\6P}VI{gǵq%?|jCxN<)jKYu" Hy>Q07B<;8|gI8\0Ǘ.`HcSBݫ5]ksa`^Qf7cikvf7!7Xܚ&`5kpƊuG \;kZȅu w\ghLf+y/ CR.K'ET!,e V`eow i%i|G_w?IWu^D)͈;1暪1GҀdh|uĴOzP٭dVQKCGOqxBI[ AcpsY u-cFjHVͥejHuM a[}5ʺ_TQUԿj/ZyQmdMVA+&I(GtaE0K1V{e Qgg"ZYA`^K6^tw >:##zꃢvXKwˬ楞R+J (5jw*$58ʼnBǔEJ%++ m G_JQ`ꅫ?S\&)UK.e` (x> t|@&@Ī/hDZ!;89nS}hJ<ӶGMHW7]3 enL߰vq\=SJB7ݥgS6;c!5p G. PI")5GX!a$ S6B1},PF,;rgHx8ZG4.!H\RRxox ZH@rB 5ܛvv'l)We*s/յ: e:aMY?կUou m⼷Kk S].TJzI+EH|֕sUILʠ)[ V)8_"8ǟ*& MB왲u,LJE,@,)0@"̂U!łfL HD1jf2fW^8ˌ+'E :Yh@G*#5QI`*tRFK-ǂaܲ$}@TtS2T *DO-fمL9jJ(3;]T&L2y(QV'n+:wUʏT}(e,q(b) )EP θρ( P0Wf`NWֽj6IQM>*"I\.}G/g6)0Yx2AZ2O7]H޿+9@leRhV`Ns&V٤:E!(ɒ @zOjKQme'$ yʋegL8r!7z\ DECi՜]M31dK@x7)@'}l[ۡK֣ uC+>]$mȚBnq3%ǵ4=meb4]lai2n<+h̾8\E:3@Od|[ PC(͇ K5Wɔhr%\,kEpp^Ǎn} Z$ GS{v:!n:9"M}8hߏF|Ƭ JބkQ87/h|o=QA*'}Xw@>P biI)r4P߮֯kB}5/㟿o~7|L'jKpǣ'ggZ=&!2z.qD)#471Wbj@S EVBV<"[]I1!YluD\4i=vo㒃1c.߰4KKy>lXlQ{# .,w*pːY Y pG6xxǛxŨy􋟾l oDZ/kY(5oRSrLԝW6}͞bS ݔGp>ő<ⷴI|?! 5'8IqUBhTϡ(j[|0aW_TBa2:x+ 7̵)e֣&[NH3& Cr<>7Z/<XZ SfͰ,\+qx§V˶KG]tjڝK#ykk"7۰ -B n\8\;ѝYhӣb4U2Mr$.rq9*eBePZ3ܱmJQ"'*(9׷ܹk]8ͭ,Sh漢2qTv*Y@ﺋkfw74!V0p7 vרpMYn>.*^AsBdzEk~ s_l۬83T-bw;֒d\mG-T&/6A9Z8;SQuZ/xܕ__&Ĕn#GU>Et;?aA )T ]ڷgT+O🮕zak˝{"&~Ud:C.1ھIߤGw3F=% }[sw/'7En"ƃ 6R=`M˂kZ5Z m>0Jf̛Hz 8^_/frnu ӎ~&Ca'.>{kˡUks6 vxu?/lءd7~]Niӊ-jsDGuyh!‘oBӲ_wŷuqӁvL1hm8VR(1mxBs*E8,8V>#@e+ yǣjE}?(b.˽ѹP;<̚-J±VRL_wwu:~ ~85~er DK4w=nGn1~ MnES (]ym~%?d=DL[%wsσ2D1C8f8/4KD8$g ,x6,Tfh'JAa%&y"`VT"BX©dR*O$o-w(R6KT} (4; T@[Zj㛂g‘IF Մs9rP{ b'C>|`VV j{M] \&Ԑp(_{xL0tk;&H}| X|Q5_ R&!QEc?w|[7O=-ߚ;>c9 ƜlZ-K ,~#{ĐTJ-QrlۭNNGe4#>ۚuo듓q7#F)Wwޣ ]\#D;rMmdr8E:1`<*8s:y5$$Tּ;>:g2LU:)LW{u UyFR[K٦sْhJ%C"aP2cRaب+I8њ/ 1 óDY ! RD sA8L e,"o)Srq:WڈYne|V ]QDGW}'$&S 9?ӎ@\][Lv!K0}&hC+FK}{9 "v[ysÒPޥ 1DeeZø癌P(QʅSE:- :Z "TQXJ^; 7ix۔^KUj6$wW%AN[ʖ݆p ʷA+9tDWvû;j$/ɂN1Ƿ >@ۘr٧2](g{!K>sNYm1<  k>qP1Gp,zٿנbdhp).% "^6s8H%S_lҴN?g`f2-p! KRZBb.p9'km#pT+ȨM`Ư:ZK*/7A ٥R*dhOuaQA폈I1euA"BJe;c}10c2V]b[;&!dF UFiW [}]6HN$6KhF3llfY&S)g _a$fIЗTqpƓcm3}|B128y>e|bW1fX0qh:%RjBo/PU{Kjۘ9ApU+GU;+0avvseY7N*Z`[lQS=:A3TRvfhscq;rMU 伮6"%` r+}mxz0(uA9R+ID_WAF\DRH1HeE\3 ce:%W2ͨb-it2g#Ӹtfd%uddRD3ԥQAHꛓTRRfɺ T :kݯۡ$F%{YFq)wvGnyu[v=$i6BȺ9.V$ؾݫ5Bibhtb*EsH녮^/81G"A(%?;q?#/_ ;+wф((Ѫ\ձd̀  >uwD~h ؃7\JXr֯~U/ĤI%2QŨ )b I9ڲ4.!È)K2k// x }y}~g }~1p vf08^YYl(34K3Cep8L"YbMFU\!+)WXС/ bЗdh|'i^'fE~& H:vrP-hmςSC`1Ӊ!IlhYfqX$q &hdѪ 9pށzB0_ X7E>.D`B2eBԴ4IL2p$Y4J>A2;Zw-/@V0AEy<މUBe>O`G5C;8O H x@'f%Q& Ipڥ.aG=}*@մ9|wSO(Rβɳ9yayap0O죾*^:27E24Ʌ1. (ws s.D`5%F0^SĵQ Kyƣ83$B!"Οzڅ6֜Qś;k7ʑo%?f;(+NcVΨ3ɇĤHj2I.?4}Vwb;:C=p_X ` .~`,_u\VtP> fBg4s5dd4XFMdt,9"ieL2 S rCjac=jQJRLsuwN{%i vȅ ִ'Ӥ8,vLmI۱ H<ٌu#3v.B3{=;Inj{{霼7g9t +%|Ja/qپxcwETDDg\\)!ruyPKIE,cR)S:ՠb " 7th#;Ѷ82%c_cBQ), \]{F`EA|Q9k+|bݫB)+LH0y&%Bv~10/2VIdg`"<;E®asH͡J|m7`%a5m[Qu`=⠷m5T޶uU-GF/0@K,0!|UQp? +oPݶx c< .IOeb%JUsEz?ƌvi+lX>c<"2')veyAͩb֖dk'pI-^ՑZd"J|qWNѼӞ3TRAs!퉝-]'v9tutAӥUϨx\P77_6L;]΀1ht9ZVP!u3^>F?T ?^+LG$5b|,"ǞB5xGT9]1l]j5GD”SϯڕqڕϯKߩvU fϬf7wj# ]bb~ 9X1`xqz uVNK)T0ev,]W۞o`%}0|N-Ӿqo94[2`~qh%z))EsxT!MhEt  |F\'0`u-C"jpM<pQ1LzqJJ;gp&ԂRL^v&8Hҁ$+Dd_Q ̊ő"8+}~kB--δX42xV&~rn[ճǂ ,놅[G&sQiGY8׀򟍢4=c)0ԳH2vvPY*4wY˥sO$ e0/o~sddWz>-90O7bdMWWv-~ ~u2sd]=\kf-|n40bs (7ݹƌnKW peo~.Fҭ?o> 3s3yP2Pwq~iyw6 n8ݬ>8L^V,q]s~ܡE++K,f: 0NT/U\ȯͥ+Jg;7,:^0g[=_NnFɧӛgX҆\LMd7o0;(^Fa ?N]?Ũލb]LƯ&ۜ+ܥMLxsSm]1$'MWX&>eSb:}:׾\,Y8#?O4OHO?SfdɋqF/gכ<81W,-\Ŋ~llֵ̖TxvVydUJ]LoLbEIL1I#,u$>tϚn0 `K| 0`}9G ;.^NO*+lyvpٛ-xb{]$ZE.Sنzjקy^_)Yf~H>̲KLV.-ؖpC7 koq7 QNح!HL+mH /H}6Fd$HdC{8)ytW R7hٸN$!UYI, *V;( %7dJ\m'0anFw.uc rQsBpg>Ͻ1&Ni5ЯTb9ǖx,ZM ųwwNhN= j&@Zy"|VPiI3k+A -4BUh d.T~MRLġ4+g5qppuf4=XMäƷ>_w}y#Ӌ@h2UEˌ'lXi}~Td a?3LfUf 2Fs=t{ݍ 6V`W&B7&;kI͓j/M11{TY8ID8hKi*dܧ9ϸowb1BWdd6e|@$`(6c# 4Aҡ⹥˛ Xl/% U5NS!A +~ *!+&ϵ~G2zFs7~/!~ĢmĞY Fq1OIjH{1FD/*rhklRPV)xiJ.~TvC]P?=~h]u3_l !k ΃uGQ&f)P<20u!D[ /LfA15Ejkf3BH=+tŭᦼ*0&+ޕO`d! ְ"Y cfhpF$br t1 %愑L9aǮN!D8A\,Z+ِ Yچu6dgCߛ?&|ɜP+1] r q(c)zY$'u"rySeW72va)7Y)+=Ʋkw66/5kĆd)s7=^\6d)*ӷQM7?fFsL1JO[8f́j;su$21K9ƢuoKy4LMUDbKv,$=b;Fm /s9WAK!ix,k!!N!6a$Զ/?bŌ}T[ pU >N>G i[nnh{A*{,Y7V*f }{;s}~Ob=}?]. 1| on`U׷jhGVvn8H.җTZjsI!Zp ZeT8nIqT(Huzbֵi<+"8F-"}I2[_qD 6YulsW f(&76.ޭ=zh0]/`3[tE)[hLB9n4f猬!"ckUNX+l-X<5U Mv~( ^u;# = {x@K3bRJ//[9Tڙj`UgN|Mp k|Auk Lh+I>0 ! 3\d@Fa\FAtx%@nlI@:;vGdeT? AN[t~:x:pa'pT:z%"&gdeS).M˻οU݅-&hK7Wiyo1FTX1^C+MHQZ!Cǀ`{H p(6Wyˢg婗b2%@ĝPBhJ!H9*%$a 20 >W |>MJƋa%~?K]RnB0+P, (H !"EBq m\*R.qxfo8Sߛ6B<[o=Ff>jPUtC bmm\ y#ٵSà3oǽd|=Ń=;(W"v}ϫ]bȟb~1df؞"> ۈpdPZ32̉Vsd8[2pFZQ}&f9ސiZ.x^l1K ZxhQ13L!o$'v`F۞x~`FGn=&(XqH2kJQު*#M̮NP Wbz?ݘ4q3a0yA'Io1]oI2˝( e,+d4:#LJ3rugbe_T2Nw#[! _࣮-k :4=\r1 pEҴJLW `F }PB@'!Y8V2OKXX(ѺJBE0z+4.DEo.?``@\#4L9!j1L`oRkK ggd 8넖%t4u1lU,u0UbJ}ٞ䲕Bb O^2ޥ Nz٨w90|2Q |6/_>ϳIP`a ONn`r>O oz&yvpee7A& 옄eRe9/&AfH/]:Z:H>r;tFҘdHORSt DE1* ,,gc4E>9%?!ʨs!h+#sq"m>{E裂bHjhw3'.02<_%+hj,'x[BCIۯ+D6E;%h&_ Zo"2a}{t:a}zFE0%{.QሻjɳZoNTHZF`ę~Chb3ϴS>2oHN`ۯ KbW(<ᄷHh, CSpXƃ8kPQWy1,-z <z5#_^.˹!x2 G[go'A\,>o" \ļ7pQXL Aµ#ki4ұP'aLCn= @f9FyFɀ"02Sq¶*Fz?wQܬHQvJQUfV"8;F jTIXY4hdȊFϏ/pW6-d[Ei-Ɍ(2P03d@T4p "DQj$y0=&hX!T",r `gGi(oo> W$)wFenV#(mw]X陜9egK [糰l;JYW\>#4;Gƌ8 s 3V_ݼdS„>鯼wSQBPa`^} 21BhL-x9206Đ#iYz`;I`x1 T$ín+b~o7YϮW)au&q,奕Elt1 3t? yꟶ~z[LBF/~67=,{cڽ\'zvE\鲄;^Y4A(!thqx>np.w.fSPZȒ١6 u3 FRTNx|) X vP :(]QB[gTb% 925ܹ%vssftӰZ/` 1.;l wPr_~_G^Ǽ vDD4:~]w7x.ټN֮Ww~(WbGmzw<+|/mDwGcU#|;Ls6r2~Siޫ|a;YkPIC>sҩsތ3uۺaóVz&d53*2U h^Є2N`EV+>HR&Td#M%EIvNVrV=r֗1NQC4NjC}Ii f <" ZI`us+˄y28rڢ,K {+(u@E8If!HM.J/l& D鱺o{McH&AMwI&p8?rad|b2143W#%[$>Bbձz& ^&gbe |T<|*:\8{CO/],>ݸ⟫h;_[exY?}R~XN*C?~R,Q$?ᱷyn 񻿒A`B/\D5?FO~W7w?*~# :Og.py\51pw߬v4`Z.FrWk򒚊bu5d)aKYKKosmW bj~JsG0V 9^ M}W%tDw|Lj>2.%YAiL4>߱oDlEӐmM ~3oިNƾӋE`ALЋUVکzu#IyY! u9 `()F&ʼnh@\ilvB BifX_'4ȸ`t%fVuynreYXdڻrEb׾p,ihJB(NlLfzrM:SP"Fȭ1 o+_rLX`<@"KYIbŭgʠAՆJPrf& s/7FI- eI.\@]@aɍ5Br;(Ṇ#G0 g(xG`- t%ri &6Tȅᐜ&RVs|I-ug`|ĶM`v,=!߷R7.5pO'<$K'(z7+*oxNj_NE\^ȹ uu"bA{Ns6uIe4YNZ YYd`ȋ D"rdFWJƂg'Pm^gdHt': ݻ~`UŮjn̙py4}`414#`l@g]2FQ`ajƑ_QKRG _\qM*99.cmfFcI&=P5 >yfHF |8rBb{Z)Ǟ+!.T˼w?o7>< _wm@kp̾&3DΘangچTDìn6Uo񇕱a_?޾zԧܒpgf+?{e|@ nXwJ[}?a\klVٞvC?T{F<" @u>+aÝ=z bl^ٵݟf?^QBEϪA^-뫫ъ[>ߠԩ]z q <UD <}D2!ř_֤nPu~llw.o~ˊG?ƾ] {{/1KA217SZ }b) }c:]tEd\ʌ]>$9/V޵D6fU 'v\*=o֏i]z>{p~WU_EyOEpV"bXȵ?=8d<^*z lLc5X6&fd8 \aqnw+Pa#,A˻ݲr+mzg콲ۆ_(N0T  dT_0p}T&1qiuvܕ7-3):UYݑ-SIi$`qRʕVnʌP \Iy59jz@ BpF9q)"5h3J0JNm&S7yIdQ,e7KRL"=/ʕRvK9mͿU `oj-sHB>Ifg Ι/@[(0/(f`N J9J)JZa474!rC2;WbLG/CN e&|؍t`U6߷?:Q ah %76vru- I p)sۅ rX(EvR @H2kbCҲn6dvq- aT?v%} ԁ^nz ?)+jAK .R).` ì(a@RS K KB$DhƠ ajus!DH~"ֵ※PDX;sɘCވt'+7֕9pnf 30D2XW'im|NuuɶXRkXYkWT6^{o['mw]<EͫսR\ٛ.7Ya7nK~rϯ^7onSE厡Ҭz3bkQ )Tu:~zHO?m^P]iSɵԠ"iȟ\ECtԎЦ%[7YGn2Q1X=B\Q[0-U4DY7k[ |T't.m,![0-UNQɩгC 0}}trVX7#PHIJ8˕f\ijo3^Z*.sΜc_b LHBb m( PxB'ˣ۠OUz2gr,l;8=8s8WW_f_톾rjp9A<ң鵘O€à-^ ̷Eu@<(CdkJ^׏:[-7գ>VA!d[|qvZ߻p 1-"s|[<2 ޳. x kWn9!uDTܺP@L1PRZr*W­ɋN"P&-hҲYzܝe0-Bq7YZC$cF5(v9e7|XV8XRq_}X17ى1Sl|QK{8Ŧ!S^sK$EKR|2qb&I*?I%x=<\XKoĪGXIbBB8$6vlm 7#@ ޡ;;Psb"w%0,, v[7wm& PY{NBҊ9dsT< 0̠ҫݭ=ҫȈu珨D`Hhe]pss%%\YCW8' ))*"/MIsC$4 * LHW!2 P,-K~Z֍) j1]sgjÀT{0H:jPώwA\LZcr-SГy8'Z#Ð޻{%+o*`V%sfX/L W1)r%t;$R\\gs[5j\e֝uGFћP[8նp~7p]1 dXݰO1i'^ŇUEqoSšTz>֤QhHK[Yn}oSk YTH@mIPk6Ճe[ƤeI"Fz$2 Xl5QaտI .}ȩŽ H'Ɯ4* >$sgfmnoK}v)!m͊a 5i UXye]~wҢe>uBSeʒ`Xn׋,nb@.'c('n-v^ h\^&`x$:Cp<1]~f?z,5h5Ns_O猵rOdϡ82Z NvR>\w0-k1/aUd>:S1'T$?xޛ=v%HS,TnQnhhLS<u8}W],y`)^ hU[s#2- lICsG(VB1q;^Fe%fr>t>ky|o2- kXw`= A#(b<ִYiqܠIs.*cTX4BZDRyqBƊO**!Rj*q)E'IfƉHO*sYv5"0r<N*% M)!m Iedi5FlGi!Oi91Xa]ŧ&e-OcE**-0,_RQa 0TBK( RB`# "FMJ*5HS8*QQN„ +\K$jIH$ƫ-ۑsTZ\ev0)!Fq;.?W ;3re/ O<'siA7i@,W8tc0WEʜfD&P VD *WyF%]r*0%Vv*4E bmEI=ENg0N֎ڮR8YzjU)lia-*{ gc7vO_*EA 'zs}dw:1 Hkwz2EXJbJ-i&JXCe-:9i%"̪E RHG@5V^\ir884Gd &MќSBi!d/׿BU:V:Bq#\r)5Fdw'9iBbLJmy!R2]"#d +e+[%JicV6pS3?X?gQ;?Zι9,{iU/Kx/ac"N鲗[.OOvpH οߖc`@݀;% Å&ӃNFpRN$wB19>V+YoQjxçu ؟˺br9k?s G[s[u!tw;k޾zݶJHj ue}H8d`! f#?O W(3I-ԮǞ)`ΕHW.y&+kTP"g/J `@ %{20 -P9eMatLeM/QgI@À+,Qaƣ: QHOk x~( B $Uc-1WzQ+mdž$+j^F%vl_~#)iA -Ze{07c܌}%ic/ |-Am3m0č0ĉB$?]xq#¹Uڕb2ɪ,o]Rx2< o?YZΖN 0lHޤ~fHMҮ7OOez$ziV7 Z2*XV'vg$mfW& NPjg[GnvncVA FN\&p`7 |T RT9+S8g^ d\8DhkOf({nSezG aXFљe1u:  ؋a넒 TpS_b$8[ ]^Wޭ'޿0@v)650!K bHfb|c h_E) (;F{ӿ( f|f#{vbْimg }MA%SM,l,"hIPm"n9{fi5ËS).UJK"ǠU":ָ&޴bz8dbYNf7OC!%R}$a[%ENWGkc( J_|=4_=XGϰd0t ̸5 Q es2˫fXNEMU'yUnRXs0!`Íw AJߌV#Q Nsg9pH2 3ٝq;tb0`C  $&.QC.J%&`zoP:W**S* ;6 D+FMQY㸰HHTRZW(c[nB\'W[F4ʝ?"* ĆI7 #Pydg|)M09]IJ~[t8R QaelnRbl]kŜW[Û2bm 4J FI*b|&X87ӞY_"q~0l"$ N0o9$R:HH'0ty b1J?ƝnrMr0^6ȪkχeH6|9x])_yf>3kĭ䍲<N>dI"w'c0yԝUW`BR0~&C9OT ;>8YT~g"=QJmMį EL eWvYaY+59Y/d%(`&6y( ;W *$ ѭeJ2Js}W_T6}ePaq ~9M(^! nwZ108wt'߳YռK/T@Ir'!pKض' Lg1U&T@t"n4d/ݬߙ$;۴wUEOsw=񧋞J= T a?`\0Qw.s*5kNlS;UTft$E_#-+ȴ|.޶G1W*( 8uystݑ?]RݩwyͷJC;ҋTYK l Qwu[-;yIDu+Q$UG@ R 1N#SSv}="FM*BKHh e- &Ri\QSqsShe 곹У|-N 69Vg%b6m`ܴE^DKZsBGu"7[e5'֞ݣe-ڢe=^4ڡq!VWu]- YW;Յ.%;U v uyZ6!֯@p:2B !#uZv]x4m3+)j$At tXtnKEPg@Nk ߔ_QSS5@J\*~dN J1iMwOje%VfN4r08)>!L-N9ذ|"]WT(~;Yz*Rc4A)` O)h\(FE)  IElHqq~r5"գ?/=2ɝX_LƯ|NXmLmV3x>v\lه!(g?hGxM27,t7[tNFNt>˱A2'd`9mwcGQa@X:+?'jfѧv:ɊUhQ1-[(N f-7ZXу;&U\*_bݤP&j7)TɇjyFT*pޱ (m z)0IlH=-,eoTJ|A:TAV5K=!E<$0Uc.DVi uMa8RE(f YIJVvg?TWH嚳z&7I%{&%SlzN2m1&ϕ0Lf޾qKWϯlяS\6 X'6x&$V`N-n&|[OvN׌LN7\&jnQM[=}qw\0P3pk9ɳ(<挍cr,S&JTW>Xmϵ)2sc6 kSFU<]0wyN/8BPBB`Xfl\S(m-A6V KErҔ86e-N 'ൌ<(Xfl\1&DmFŤ ߤAo}ΊNkV,&) H1!7/&;ݰbItbek:~soW79Dkum5ғ՛s =R/N<5H*vӸSpΊD~=O%0exPQC̝d` }N^ 틋(8Ш8&PgX朵Jyf#رB0T@\:#2V [._[No$VnuJw$9tK!_g:1ƣ &7j &; [ݾBZֺvr!E]=Op*GU!KYX՜pmTv0e/)QbcX* :=0Aw^ Z3J) /{R%|=L(, gg.OxK%}<|H?4LnWDe5x0J:V9ÕmM\=]<ٝ!ch7|tc7hPu9WK N7fF?0nG&wal9V:@x 暘-^?L}t}i'ބ,_k?7zϒq^g_~S.'RѼϾf=+k̇0x#gwtgm@4:&(J3VA9K&x/  [TϦ%_`{tZ?e:TOsȵR7~T2̋ZYfMlbagEmv¨ |upm5{o?d1db=~6uKʂMɿ1+d㴀cVrT̷ e- P#'EVSfCX"vHQFbSY4thI Gg%*@ vjI项\/ _0O)yp6W1]q ڔɿ~vY=ㅛr% $wٻm$UnC&w`slr]zvێeϤwddJ;iKj^,X,Ʃl&oU#U(I ?< ^b75 a޲VK{Prx)vsO^oG sDbٴ,x|&fR ѝ&b? !-_'1S ! >ɗ ߬7oO̥29p,F04 SZ0gæ)6qiF&b< o V!>?*ځô8 :a粟'8IS.>k#$Z[4cL6~?y @FppUT ]nb\n*,:m z`|˜\PgBD0{1gt"cXWno^-W`ٿ_+Dk_TI5HW)X-Z[?\S&~9ԥmwۯY̚S@ A,!OM>56A?,nVǩBdznYmW͗zGx}gwozw0-Is:[ {ysmD ֶ!>wapL/2hT)O>QlRW6MJ84Η1тϴ*NGkaLW({*[[a#!:L޿R 9eN.9aaXj_:̔\R 8o/x&Sݿ誄:̡TePwzYt'<̭_}]B>:[/iUmI]-&$V/aR6Z T`u ת@G=1Hps(u)RKe0u VY؟ΐS%ĥIj'N[-DXYk?Vެ9KzñUVp"c/(Iʥ/P7Qh8QdF0kDc aIMS')>>G \ԪƏKA2l cE3\*$-}ݙmY]Y)npwV˛ng):.6R.Am\H o%T)97xp6Ղ Τ:TSO;p>Ռ«ϣ::t>\9IK'͒ja0a\IM)؊)s o8YRp敦ݦZq{0l&Ob<EgpCzYߗv6e_NR{?ur d1JˌٹΉ-~-sq!\ӉˋKa> TptzXsz? s&O\-ٰ?{S,E3Gi=TXr /Dl*>XcǻiʯûEʠDw~jݝwewkBp)F$M/bePtbQǻ]0[4U[6BsPM zXT1wnG),;nьVnMXnQ68wWbePtbQǻ]0o :E3Zջ5a!_ؔBvF%[ *N;jy˭oN\j-ޭ Mt˦DEN B4UDRG[!{on-Z.RVCI `,o~((~f䌟d6X7?-k`0_own3-+"99\`v1cLS]ћ% ])ؗW[9v1Mxܭעd2KQ] 6Ya@x}3V>,o1ʛjH#$uLRH ڻI꘤xPK|P^H*< kKRK*+ޢlEaDKoz㉱axb)7t?Bㄪmj<$LozQkgWb=#=Kk! <ʰT*/o}ޮԢ7Om 2qE?|W~}]|Ϳ'J?RRty.9{̞*WxԻax߽;<\-ng/y#qg k8J믿U_~@M|~6ۢ_ŌS3x ?0 >9%?G{x^>0-\0f/_y[(B 5Bo,b|,lu}1k$# w_˃% 7 j\?"ЮAZӛ*KKmUNI ctRZl$p;RF72m$r>S- /:ۃ &MsnJ*N)B`g.-F0>|Fd!SL^t6xR.m,k=Bf^4[9̧{)so/**w2#_ۓ["i9_%_vdOJEo6aۆ20Wp 1̊,RHR"Љd25,32BS,2Ci u:\ c9c05i&eD=8"3+>^N R߾-qBe]Jihi84_ōYD (->-K̈́?&%)libi4 \'cp]"n3-Gc 9#&WL r{L.ĕGUGuT\/PÁAFHlF1ge?8._mNFC|sI@+6:v 1N!0ThG`ZS'*L$Y.im/~639oMW'x*-טJ0}n;X _M ֑ٜ`- VQ䈧kuJcɳ$!֏,SPEJ4U!!U%*]SQ9uYC3xZ~Հ: Ax#ϗK=Eըh@d*>LQV7)BOr Ude>_?{I+nkU{ZIWN*b>]ü\ Jju_Fr9B!IO&!c-^,`o⭸=ۻv Ȏ&XQ1&/g"g}C2M(@ eo^7/Fr/ h lWaȘtjK`8k୽7=waZj"Hp`UF5ّ"ŔΘ)Zi/jK+K{fhKozpּʍ((n2J/uwce2#q~f-'o N .R&m9r"0Ũ UFp$0i*Az{WP_"P*N0  7L{}/7e|oG/y=jRu0 /@&J@:nfv{4@k*Xo ba.FA/6স3L}kv_U,ZetWþLMQIWRvV^[cќHHK_Egtc@  bJnTxɖFF ٩jE"ܶ y((?[U>:>¯ڴ Q:@nҀf|Np>"N[m~,Z!tC߽ }قm&1YŪ%Vd R'`5A$.:H1mW.~>K)h|9& D ~F6`r%ۗlp,[t|N@sOu䟛Gv>?aIe8Bgn>Z% m_/1LY`-1 z6>G8q>GL V%ޜO5%_A/Oeao њ^g;f04nny2{ N9Y7 E-Ǵlo?&E 1׆ݎ6$.wr>KB*KPX(Y1ef1 B=N0'V*ZvQułwɪӬa2'G4o=7=EmTDCg6:I+ֵ29`<·>t0F%L'dzmG 3I%㭅gu5e|f$" `Zˎ@?[\CϳP.#׎hb+^6`AfO$դ尪9ɥ"ݤBh?y7{$\"a8*-d͙TJv'E\>dÅe_ꏆtkR!=[NNvLu]ZD &ZW@39FX`t1/ i?"ݡ:ΨA$ Jڈՠ[3)p+.>kU'beӴUW5DE:S_[;_>gO's/GJ58Pf9B0-i@jDŲg%eэ#Xr>ykP侖'7 m_0fuцk6FXjZ`"Bm-dCM+WR&jҧ|Fʖ#͛Ҕ߇ +qX4=7L0r {涍d _vSs҇R/b;9OSE4ZSoI1->K #_gۧ4حܥ(D::DcU%wܬF쑣н#e LK2lfS`Nzj,;ό?X(*@te[pun[iTkZZ@:'jItc bm0AJHmxMYUL= (R4zLs4eJX}xOE(c&X\bKOv59 g$ݳ 0d~0㤇n1v0P^&Ae )}jQeeXe@  R- o25Jx VXqR]3kgu;\B; 5[Js9N09<{ݣ`ߍ -ࡧhq۶P짂ՠʼnxb_Wj 7kwCٌ5p {nѝ [Kn[S {4;=W?bQ4\ZW4t>$`އƷSmG_C*7ViplC)S?^L4jl%- Flg҈5SHޜ9B(~wH세i&:V*z, cqn+S!+d;rU&9b4X|E #^i[?zwc!E`2c)$K"hϊ)VBy].DݿY;JSJ^[}x d&Bp9R[i( z`{8;tDJr0aVJj'pÂbhCi+Co2M,z"UU1EwStT4Is_xP~O-P[D+\Wa1r> 6ŸX[fO$hbwҡI<p =ĖLHA͆Ss%唠ɺPX18V+v>QXJi$J5FPl ٩c?ՀI@/u"{'%cchĂY.j>5l TԀZhYo7EڊAڬ_oKW/Ipj@!VQHB,)A*W%RC o[f|I&EQ3^껝n'\.K5]Eq'걗g2[pV7 w$C^-wjST/z_@]"_@jEf7D)ˁO;jMOɝwf Վc: 02$Q7v./\ݯF,xq=2 !Tw@i_-VKR_/߿~=pablK.X~ R-X?\]P.K5e@_Jw >PD]4LK_jS ~s%lb('tF@u ǐ>ftr`zH&>spA$Mz)a(l@$Qv3!VuV5T$ /jONpPu"~VVʪ#V (uE~O&o,ѮUiA*wmWr 꺘:K /R2@sPSDPvʆb%'lK|L9AacD7[!>[1Y d&Kc9bdlSd+k)է"={ `Z&I |r=&;fgyA0[NO} 8Nsɝs+n7K2n%iro4O&z$E:MfC9 c+T ĶUv n-3I,wejb%GL=̇@pK+*́q~*fΤ4{?X_-b=pq꘡c Y2Ud_PmG8f=v˳u47k C2 EΎ2RL7MROYz֤0Z(~e@[y{io- $N~VR|Я-aJ^[̓91dc2Rysm-uo> ':(9zD2)[/=MOm\~|f|Z]rSXj+i?)s&ewl{ lmiA } ])co37HCIʄz;p8{ЊR-Z0_ ACL PmM_itX(Eh=u8KA."L'K~.Ca_{Ԥ8Z.Jnֈ+<4#Լz/Mt,+a_Fd><(*%e-$^뙓xП=eV|3Sy)椳 i̦^hF;R=OOpӴv#?. @u:k(%Ւ fD}/}X^c0p~:(n:V$KS" G>2p#." . cP=apje|0y~p+ZM\6XOhNTX)7W?pY7TF!PAezRb-zy'R;_yћp2A:[h^ ?*'.IX#=~^獖-7Z>oY< Po/,ycwv1W$qHPz{4 0R^36o"BO(IT!&nS:`cLfyk]EUMg-MsSRЗgo1RGZ~F}s7< {-fTjbsXQBJ-mk=;5k׮5ļ  3/+Ic|jP-cC .:+{P'Q(V]I$ڋ:Z,.yRSza`ϛ5 iU.Zeھcc^臇(.ŜH)(u9hgP췱 qsdS2W5n<ӣ# /2<p'Z )uQxq*]mNmWS_-NߺI\(IZzs$g9krZ/ kDU>~$%aIwNKgIRёÍ&aY^+X>!N3PYdMiaψU?ȥ1h³<&?uCQo<{fGmzǮ5 Ykd][%h4o!o͂>ԕHE>{S{o_uC!!md{aՔbÌM(J#nr<&ƫI@rZhLˇc=9;-ǡs# 6[a7MizDM[+ ALf̾X|A-PB W"ʪuR6G87̸lW<^ 0R r1 OC&E=E cEpՉ-p,eV<wǓ?0j3g%*sI[ V.);u^ЉeԸM.^M鯴áRz$ uWxH(y jAQRn^Q{',|tՍqBz0-r u}P=,xokhgSs1o{ǡޅZRNoQ۬_a%+j7̂v m9p1u?< 19+ʽ-c\##ö6l8+k3A7Ǻwt67*H- pc%0W =U bLV{qo3cike>xmmf<;o3(t7G+ekeIBe8.=J7 D'5d2Ύa>yX5_hڶm}4x>kYcxǷ=xFO9źa~;fz7qqܗM/AHnnm5s =پڲd@SVU1IBKLz -v*m8bs`[3+^Vŀm}!lZn5Սb%tayr3Ɏbu"Ţݥ>C܋yfOBK̭%S0o/V,۟2C$ٳI#!.4|;DfM x0 <ϏdqIJ/]<"/z0$X`feXe "+6m6minfYEDVUƕqԒ(P: :(pDbқ{&DAezȽVFV%nM^ A7&VImF v7q#:@ (f{BkF.QD*xc<9Giݙr|߯SzYW*^?=Ws@N`T÷׿?h<}B~i*#n*'{GLΩLmcmf.1YQG}?ox^Wqj:`϶Gb<5JxL:̙Έ ٧{R~hrbՎ5wR7;)|mI`z#y׷ Ydn|wʤ@z# {r1D]E Y L}sD'֣|P|5vY>mlH,AJ >r[d8v`|MoҭX t Ll  ̕2LII3<ے@KeiGwrwPbUDC-9! hCJk,XO4eL uZ1kT2U2rzk]:{iSufm_nWԜ4 ] DFDW Jgt+~6L;9pmbRUD@D-Q`pm B$2J^U6' Uj^2rB}@R GDt_Rޅ.AJ+- _)hn(_%44TBpA4^# Z[ VP H2*5P>ull Y׈=^KMt K09NNck|A InD0\g=7f0}hѴ#KXc:Χw-Eb$A>i_ʶ tGCVەܥWKw3C^ǶcVS !l#pSJ(R> UmEZr%{.Zg9aC .0+ )op%S*jJ} ٛ}ǶT_g5_%$4 '=ɁSx|n(]h_SG7X8;dD޲FI`ԙz辷}WoοҨ4_+:u4s*eF>7 &]==8jdb>Fhf'/7 ›pSͶ9܄ZF0$2zg6{;..gq~~:[;1%|K9 *gw!stģHyYL*ΖJXAz>G3:ʓɠB9_@vS%Km`­Wӿ=ץ-O\IJ]ǰhr՟# y"Z$S Vڭ8gj+~hDf{o\ՠߘ;}e>_y;L y"Z$SZN{T#sDVJ)]v#F"Zňj.$䉋hL|v3|ڭ9SF6f50݊n]HqTv=` _(09Kanp\?SQgr&Mו;tr'cLh̞1)йTdR}wwxA ^^oAD1o1 @6ޘH坪- yTu&Rʓ&~|}Gjf RoOj~G4~qP iݙ߇?{Efm2-zqyb{ϪSCTգ4D( Q=jQ]}*0jD#VZ0G4:<">kK@Pf9T_6xEz5;L>ңtWut$u;zZ{Oa;ŋ/h~˳wK#8z|DpZ^XX:"Kr,&ձ>[&Imfxgb,M/i'%yeᎬHȀvvq\FWGQ9@268 =%2qoG=J+̈?5Xz#p|=0i 9*HMEuQ(XGWrʠ[+H +.m^5#uBkhC+vNiE0ݰIKLxO3wOŸ!}T,%HPZe drOPekKTu$`X ! :p&m,䊎2x|a8"a*Rj*ҘQSR:*%^E2Qu((0Q4T^FQ4RTCQtN˃urP#> >Cn fߑ4o*_.ǫ([Z-y|^.~X[>[bexۗg@_Ũ-O^Sq7!e#N1Ȟ5jbH&5sP*wN;ɭy_+;֊2VzMW?)&zMX\xci;x5 {CvYu=PY0 z@@_Hib5L8DF82m[BkOJhPk`{v[ѬF-U R >Z==IA\-$pPqVP7 'ѕ9EkQKI釫.du!!O\Dd OMMʱY- hՕp<yڸg+a)] ;'.E2o-{ڍS24jnRF;yS>(bDt !O\D7dJgʔ~AzG(݉Z|ma{ Ӗp:j԰%75IE@H?Sౄ2U|b0z}>x&׮Qi8]&7?~x{^^<;]f"H%Kn3q̿T%d`DY.i Ȓw"d:6q)4H#}Nɛ#֟R2e`,Py氼YҩjTSv=-,ỉ{tt4JG''HQləv{m)>EDrf <3R= al=aO#'TQcFUQcTE˜G~GlQbtt(Lfn0|.` WnDzj=;F?$9Sa3K16o8f;pٖNNVesWFEO< n}ȥu&Uc_"PW2aU?^^$ש˃-pT3s7[YJsăa~ bU+q%jUThdJ{ B+.j5Z|zg] 570S6Q|Y ڱ7{cuɬ{Odig  z+޶i`'x|9Đ)|Bp =fޖglj2Xct-ԷO3F5ѡ(a+N.P &2Me(0&T'L9&56 *(g\ݮPxG]g!{1G1a[vQ!@nP)) yV\QL-Ru.N|)X% <5+h3rdIj(Od4IԽfjG+^.d~sl*/>-^?8KZ]ŧ/'/2 /k ּ!} /tk󲖾|:@^0ʺF>/k| 9#c2#|ʑRrfqpWj&L鼾9J#Gr(]5/GS: ]yВ(j atĨ PF'rt.T4tYt Q7LFɣdO%{(ٓGM{)ʔj)|k!#xĩ87Y܂Qpu٢\P(.!&S L5L197mD4|ڱZ)#e\s|49+M홄 wem$ 6X"Ɏ φ5 3 YCY lD5 HD/窱,?pxL#T_GԺ;tZHk0PS!wE[ EFKbJ־ `/Ba͈hCn5ai91#]dT^i߳Qp5-dH˧==l1(m'EcJxw<0N`K 1"% yS/,e6jXpՑq)w=0\K7"S<Z ZA{{"L].cm;y҇?^⏅p鋂q khbs0˜e#w*BxVEZK)9.؃D:z=QOELbNEyɓfKqzlμKyz*kK-p浃h1w6N+sl- B>F;yje߮+8 ebo$D yÑR-);o bҰ'Qs]jcMDh`Ό>s6Id}pRvؒl>hLGs=g{fTKbT;tٓ\4H2N'-3͝÷k#?(} y8d۴S >8 o>AzbkNm u2mnT"ETJeU*׃ru` hb8<9 6s␣98 &A cń[Rw۴}Q6m_~4+J-o](X_%{\6Pr"bSJW*b+.5h){(5u4pQH8Qi 6 -<13텦F2ŵж#UԈW>ؕ$ick4°40 S;j. 8H09Ep<1PbVqFM9b! VJQ)U` "[QV0p"7 $4Am TL.tc}}fE< m@{.mzdl&<W<5cLY  r m#A(D.aM::whuBQ<ߝƁ,b^pї{R72>9q|L) Rd  ᒢXw?;;K7{KC^Kj:d(HD֤`񕄟UXbb Q#fșNɮXcprIH%ƔitBcǁ+QUϤ8w`,Q oXfb*41v+`->$L", !4Ĉ 6%U ECi %g.6ꕨ,I\P*YP zO9CU$YPQ :|i%.Էl3!V ړ4wJwqN 5@h*'oO_B<^u2m߂jR?xHxUv3%BË$ z,!l*՘m[|?ߋwTߢ} |DYLх^lٶ?xzzo ݯ_@(rbGEJ/weu -uK1`ҙx҃y$:3< b=l3 CcTFu?YM/>@.9nfbʭi7j7'3x:-ˮi,:MD{sYqg Ktv9)6},V|fKx_gAV4sy&hYtx6eYҨ]'dz?"|5=Iԫ^p`Ѻ;.+["cgHw_fckqҐ 䅂؆՞\\~I+uq;ep8׼!TCAdsO,UJX_l^㵷.emٴ_t}9m_GGbk'ưJ]ՃG@G˵zpVGoQ阣XSnT[ʽySc^yF]=9o}Z.J3r;3<3\Ne>ֲy\OCL\-_V\ֳmR~*[2gppD'sڎbՆhL0"/YJZwo yv~"8#1"xf'ȓ,Ng5)8f` u]mP1ogG$8^cb^xso''V[]ldeYD"o8ףS8텣8o00t?to_}WR~}i{`4C>+@tWgn"*ET΋e,?@Ψ kq634r*T o( )<zgھ3m_~@KL5nNKMi);WKw>8G 8j3*dyiVyHRak)u6 sl#v>"[ TL()Xu{ mmmkc#$KKڣ3D7|!l !mG$4'\ ڣ݊FD&w$QXB젻ir{{*\ͦpN ;kAdcD4Qm=*O !tIibܲAL"/ޖRm)UUT6hBw_'34Ynepty4Q^;b]'NibMј2is}/:a٣ ,* qb;K<6L3,FO\yǠF30 ؘܮjʛVUF?L@#OAMӫm˛.@?Vׯρc+%:4Mtף+^^عN6Y>w_^3›7W%dnV 3gLs);3DX} 7\S$pŧǃ\|z[|fŧInG\΋OU&ŧ"0^7fpMݑ|Z-7 VSr/}PTHgC[4&4wI2(?rtG Xe܌⋖MAV\BF\0-t-:ow;o]ż_RpJj\dEl93FQBBEd&[kwùwC%fNT M$$-DQ3Xy|TB(M ZQq($У#M! ]P eù?h" Q-L! ^A??3ylk^98:pbG6O?_d/Ly85ydӍVG1ʉºe*KB+qZ\XZ"y$[$NE3Ybha1%w3H#Qnv`1T;G;`|:y$ GTbfZkoJoj2Eto*Ԑ Ň^6H lR~k*Ԙ3r٤qlJ,ƢXbB!M&^38MjOCD5bX60 4XN@eU AG+ <)ڪLPk42^7s(C,~ ʇ A?p٠aVcDfSъj{jIR>}kS"'8zAtǑ&ٵ5KcU:ѲM%ɐھBk 5ݾۖ7>}) ƤŲ{2XRix,׻pn=|~X}60[햒%XObj4h$8/HL۾%g-F|ahfx?.=6F_SIwWUZyH.X ED:w$ʳxnw,J[>0{g$X2sY7~)q*,=؍l!nFV/7'X2$Q8d EIѨ6#̉tPE<,lߋ]g#v7䕄 b. 8#,01bhD("$aU畣*"+E""KaYR8&Q'TcVTFE; xC`H n^4~ǃ rP<6(0hFGnwR0AWT$VD8qݙ73ز1l*iu%Rw_LWrMZ5P%Sbw1HP%-j]*H΄B%/GOFnLG!>vr/?؈$k!Y/ɧY7OwNnt Za]es֭vp}ظt(q3#, W7dSog3FH3^d+]p\%Y3vCTw`i>f*n׵%t~WRseOdHNet"SK(8rd|x{az! C lQnizwqnY[m=dAq^l۾| r`;ZFx̓Wۅ}!̑謲r:2I`icА+R:/S ˼}=[(OĘ˩ ^r}a|A f0IX,f:d݃Xs4w͆[2+`-~~AW(s%fW`Qz2^;h`-.yX< ѷTT* bZ-/$qqٽ#4ȉ@]%+Z48#V.U2;\pfœ9F~(mni65:zL;k&b&v4*yn6':L; Ҷ4NO֎WWɇp0-1-hs8-*9H̙5KOGSBxO*}. ӳ{E-)\sEacrYEQ-n [c: :U| SY7Kbmw8OP?= m. і](y}8E Y=s-W(G:'v'Ӽ=vN}@P1zsu-TC+ٸ]9{n0ځQ^蜩ˡP\;/y!op˸Hi眶߁^# t=kcXr^kymTQ:;p7N1J Jhqu̮#tJ%<݇f):7̥DpjFNN2%>7tHPPG [] AR$rS'ۦT0VRƣ"ui4=A8VVhPU! iг/V 9b`E`'`i[~\'[rm?/.+R4U.ec,?q!6kM@8 KRCcdtcɈB-Mp)mc}td)b`W+?"Z*?yH"뒖ptV*6^pb4g+Wp9.. aE4^/- Wn-L U|:("_z](0gGHҟ.A- ǹ{Snn,- ̆EBɊZZ,r,y`L%) 7S4,gSJ;'"Z-灎秡Z x,!MM_SǠ)%.0`)sAI f<l,˘3ؘZ;Xs u%n)ryQi15$C1 F,FQۚɮDP.IjvW!O̽ms%3 L$LV4×BtNwA)JL%wߢoZQ9XI=lǞڽ| k{;ɀdq[)䋁$@rV۟^3Y.e_o?׺Z6O‡//e`&(5qc% B:XlP*$(n~eAe_ tx#qQc JJ9=U~`P[(pC"XϢp G @@H 08%KaTkF E&0:XF04 .#rc[!’QMi Z! 3@`` !#)F E0:0C42] ]1$\E`T1@UJE?BhbL[eek )v &@fe F"DKA\HbD`c-Bsu hf1KdZqpD`&qdl%GoX Sy {ajYvY{ Mmc{r׹ w+]ݼww O<~U`V>vOwl/+)']<䵏l0?ue{y*-^~U9VvfcXo,?rU}j((lzou)dOٳ=LUH-E. 4J Qf2KURzc58ÍeʎgYoeREk=?<]^^UC^~=|ݯ;׵B~ޯgk _Xjh5o֣յ j߬Gk7_.VcG7j%ޮo[ Egz8_!ݑ~ '@nH⼬6cIʉ~4h83$g6 HDN]ꪮ>xW tи!vXcCXSıƆz0cCé6(1E%QTFcm!;1b1P_E-ƆzSБņI|T*~jy!̨;ctcVW7˛K^^L/`@aEjwtL*O$ѹ:o63O 0 8fhma!Xc %U LZ:_+֍@lj+VƠ+IVm nL5G>_(=iZ^{pV\b+'PBC묥9Lj7S-5Rխ3TsY5DH9[T3UC\S-h։mDܘjP :mN5cSz0 8]7$ݕi(8d?ϯD{ h}~-IGF>,f7~4O߅[͓߾\,KCBg(O }>¢| Ų7SYVF/wMzmf&,+\ B0EU VuxeIu;n @5[9rƢYC*GA3t*ڗxn ! S~wwIʠt>cvKЩzoޭ}GMB0UC]ލ㇅Cx2ƻt־Mw!GV=n2ޭ @3n&J7־w!GVRUM^x2ƻS"/3mj]ӻu#h+LQM+ދBKڲm !Bu  ! S]zHM2=nmeP:mw;d祠{nKn] ! S mĨhΈ;?9$3s`%65b*w厲.-A<})#(wuj  .P]&wN.,B=l宆cjL)6Y>G\.-14)r4WK;%fȕ\Y8fjp4CZuj TUĽ se-Wֺ75R2HՐ+kZB̡TքZui (>$Zuj L&$A+kZdhx5\Y$jj A6re-W:*k5W֎&`ëIre-WֺwfM(uZBë)FyZ^e ++k֩%H:3kBise-WֺNy9deMre-W:t!+kS+k֩%p1ʚVHZuj U$$O̕.-0ņWY*+k֩%0$WYI8+kZbp5I@+k]ZByduk +k\Y˕N-NTP+k%'T1`[(kOߝ* mm scM3a!6 pJ'=w"͵Y $~Z/"R(kPj#NoED@XiJ|f[!a1OLo'qY>L!`m˰8O,]yx`vՒ??E9~i^n-'XN~$PK5 Ŵr`K!BQa"0Gt S,7JN"FfH~ w@nfl|OS#HVW7^/V__R&VqT0dQ@|*NKivNǭR֚*O`a0 z* kD2 8h`G¨B} 2\ 0p'B0 2LIv__l~5ZO_]@H2Z@ K.$ʁWȪ g7GT\h1p66דٿc (z> ˷Qz!y1_ނn-* ,'"fz;GL!K\'*>*#lzmiߕ)!^/!o}qoc׳KSr?/Apfwq0>?s7W%g[0~!N|okq™7oE8ge{v3 Q5j l` Of󿂻I<e<|v ojLJW5+{I9|a Q\,ݻūM\/"N&SZFJ\26Q LRxZ|_z\K}z- cOG=eu&9kz\hDFmf^z5;KcDBC$!D?ʐ5'@|ntQ&fMGXJs}[:EzE#AWFo)B[d~Np^p|Fcu*Q~Q!7lt"kIÕ~A68fE>x`G2zzq v|axbO!:Bja^yQa?:[%4c.8k3^8‘9P\A#$Js'E9NJz2: 4x*iŒ=g^EË:`8`TؖD; WxRP)! Y׬n%HT0BUI5ܐ/)wZJeΣl|"e h~-:+EEwѹk2;g:8e {j24#$ֽVK[{.ZŠbVδ`l8F՘U:jOEvГfbEvkkʬ K AU.MwC]Cq1ĝ FH)vKǸ`:#iʉ!-@SA:j0H{ƍQ`JR~v0>dӚ,d, ).&}QG'쐾H!I3D#M5?}ˆDT9zN*VPn8:a}gZ2$y뼠sRO0~wTtvAxA(Np8R9O""4II+bZح*B=xT~lY,7q!!t&|hM|GU^~[ݎY9v>![qϛC& "MMގvQϡm@/K۷WN eVRD9 rۃi[n_nLjD0B0dM(/\Hrp ;Bs2{2 JBD9ExQeJ 6Flkew; 3!xϡ(#ٸ{flq{!p6˟n>QZErm082DzqPnLRg[m ?Ǯ}n>;D |y^( l\;IYaA {(=ؾE|}rl=[v"}N%u9_\gE4#Gf1T4reLE;lmr a ?l0}v.JG3[׉H[xYfߌT1uaAv.~欳-byckkTh[YwV G,6F4CM@Pߴ=Z!(-ږi2 kbG44}s!4LG:; u\r$S2ƳK4ÇM_UR꥾qV ʏI HȍV݃d*ʨE>i䅨MlV"Ed}W6縲T{pT}vC~34B vZ'Cj!:Mۂ+gm"RL0iG٠PҊ=Tȉ3N< &F5&ny1sZ"!R"(%KisBzʲҏf^^Qi l<~ZMGf nw9v`G N {+~up{$-\ʹw-k׍YC)TE ,V>]&Î "z`jS! /M+zb})?l#ۜ߯_m>+_ ^"vREع&,"6F.ܸօw-A $XNo?~B]~zZ IN~'x_^W]JZMbfC(w?G\WrP=w9^W܄tQmp,?w~>B \kV6\H).K< S;P;+?|B5Ⱦ3^/Hf<%ܺtK!,!1`ڹ1/ae0ܦ:g%n;|N&'? jd-WˣG94ꁤz3]J4YNPt䜀I8O݆wN][˻=7p(I~]ZY$:uYjkB s( X+MfeAj-l]pYT;A^^kd逭~"1B"9h!ju'^. ivfhzD4KQܗv!2mKǟo#ÓF'jze%R4I1;Lޢ&81Ɂ,2YJ>Gϻ$ވ{stݼLSmoÎѻyB֑GVl麧Q< /g/_‡jćj2BT yR-I'r-&BNIc/[05cOh {#C_4C_ގ?=ćRzv(K=GqR=QTdѣX2zfS2Dp$zQ\2GUJSkѰLƋڡ}cdSbu][? -x{u`Oͱֺb4n^jpuؓPskmdo׍ DVkQ;Snn,֍%Z!}F`k d9(#g_yvy!q0L&qTWl 2+aAxUCRQX _~kJsuuS}LNOg 9՟>9^ZK?dz#% ;/i+].BB66a ZYHNسuXEowpXUd.guiz&A].2?ګ-zȃ` \pvAKt4Fa=Ns6{з -pۑAJz]pЯ&_=jl 05_FRPo6g,eBmRBBe 7;$+!d/JThJ$=`EI\ҋ٤6]:]T& BD-\_R(ڂdQڔn14:M նÆK+ˬ1V8.9Frm(kՏ\eH>Ғ  =$$v/?N!fx: ::ot^iz5ϓf~][ti٭7g7jADΩ,: ۄ\/'W 蜙Ϧ݅naSz@dF :UHh;[:cw|fTkmqm79Μos5G<鿮/?rysw^^֢'ɐI ۰u.߰JkYeQ=䯎O%9G v9GxC1Y[ I4T`!?^:￿6m͊[ރ:(F{G%Gjڹ5\8'ST?^n̵ܤ)?KPsLg?\1n{l=8Ȏcz{3G"q'L0O@aqvķmbd36_m&^1.Qbkoaԩd̠rXu[ F%=mUa~cduzsWF> 7he/OΦÖŲ0> eTv䍓o[ÑrPX$a1[󠝾}}wO !{rKvNȱFi(jaMK>J>km Ēv ?^Pi''''\c)f c؟ J*RvJhu΂0䊱إ c8O2ٸ5tcvhL;!҆τ$Ed^Zs(c(#Ź"Mم("4C2p( 9 M-gT-lJQ,X 0(vH9DْT(ͪU65vY#((@@b=hmR%[Eʹ,"ȜWz% ջi+`kP6Xolg_jF@ dC<2cY脊 PLV$%].68B% boGa "،x#DߛHE ;+?.ኃ𚐔:0;6>RqOfgm+Ҝz"B__XB8,^>J{KG_U~%yjIѽHic0#A&_UEq_c;},IRRj []* tNGoՇh?ߓ{z teI]<,c}JJ̶]Xrp~lцEq~fooF7Q fUkfUo%[޳5}cZR7H BɏN=>(lH>&dJCLn,rc.ymFN8g !)vӄBpb:m%֜ҕE׬\5 ?_4(%FE妶5vm/ _GOgN8˘T q>u4"_\!J1 \UxҭnCC$Մop39%jf k? Ef)2LUiQs5j>)#51,jEL5[IaˆԔ)0⧚my_v;5(˴QU%#%k5yԴ[jZQs(񵧦uj#?ܒ͌Ɗ92Y Eb,$%#Ol*uN΅ī6t4@^7>3IaHfjX! 8H3lfVJ bVO%h U,-VȠɠYm)B96(['0.?_5"R -tbhF$(Mtخ!\5!"y[pѡɍLF0GljKS2&%tI(8@'mhEO#C}@L%  ~ 8"NSӇb$)XNm1bV 0L\߹v+`@"vj~|Hܧr ] Z+N>r~pd?WK8H=p$|b<<7#4;" 뵫#-l M(.%xY/^o>jK6MeA hRU 7S(BфynZyquqimU 5Y͙i\ ~>v#aR͗ݿ+uLȏ-WK;ᓝ5ӰD49DVA*kS`'X\[08{XPjz*NBp!>=Z4g@7{ WHfHL$)V%Oοi/P@竒ȒO`E9 ^.ѻݠtޤ7̘6iQVY ތbgqǛJn7`Mw(&FrW4 |||-ڨ H?X.iEAB2 wW<0JuW4>=bp=k7&B%D$HIsIT͋d6&h)0fm=''vսVsDJGıU^08]ߗy|~cyK8ګ_2S)|c}e'h6Pv2a⠅8xV=ΰfСXht| oldz <?5P1 ?ׅM5q?@i9gKqZ]d~q~ǣfrvCuᆜڵ/1q1wv3Eeh^ِ;NpJuMz#4cΝ~yf@k& m$vʘA#z)rm!(lp8 S#Ej7$)ŗl( Ի, s)-CEzMYJ7~A\"D-B`ja(AQZآ;c[\6O1^r:*Jɝ ׳e=yQjbyYl#(k,1M;7^|?]\zurOjR24i6DbUN,Z*;̫s"5ZM;=;>w<,e*w<Q]œ J ;Sk]݈LI{ԛHSo{q~\uAoUU<.dq(ѢqPZG)pk`\MXXiE7AuC>; XCόv/@Hj`S,>ǐ),RUm UK6*vS,Y~!ʉÿd]wAߕe~l^ )piCfSyQdjks".]CF|~*8t\K.t96L~b5-57zp5֋-֐zv }Gⴈm)\m= 9ù9+E.ѡ!jX'4^yTdri% $Xz8-"p.elKaLG]vpdkmbWT8o;:c?Ug'P9G̎$mN'bˋs%VX)|wI# gSi~_phK3p_ԯ4 PƖtP>Te4`Tgq,/eƤMRxm,D3j+qR[aCIn iKKc&xrJ88d.xo>4dcL0~zInq Í9^DY==8xU!`4fchAAhđ8yh芁;۫ ϫ 5d'{p@G@|*4a20TZ&S8o9K2%CIcs$9Zڌi R"7C3=x9j\p.4ܓ;Vr c01Iԡ ĩO>ID7,k'EO+URͶt2pFHYVi;, P\UO{RI_3D}<\T&?hA]R5doO/$,QN2&I:唔$g"vqLR|I4VA+)(BRE4S^|5(y$=x̘s IN7ȥRK{ v 2#AЃ26XfZ8 vn$)H\t<[1.{lbV1`S!CVCmn[rfIUb(u%$7bi+ڶ/(3w c/Gқ x6y]V/)l\]^\2.b51tCujFO\nJL>_LnIz뾼jAdP6o!a/cV4]dÉ'.$ᆑ9F}f22kCT:NV˄y(k3M;soAh $W*>P.SO!jw4M/m1uZ6HSr+riy&$Α-vedXӁG%enFZ0IIv:H"=9[2m$L9:KКgS6mXNO@*#ZOd<X:yKMrb5it \3=9&K%K ,1CzX*-b/ӫW[crOaBIǬݼE0#1Bs3'夕|R=)M_%g:0]7>so0i@{gk9$x>g+y2RAvSai]U$ad” KqIRA &$I 9Ҥ£%`KhL3{T 4Oo[fQFQ4 o6R$$iXiMN Y9o,"Ky*晠!TkSOadR4!Itc\e2O8&gZ2έ֞Jh'e8݈JݚUaSdL$K(eSԜ̲%VH pT&'*4ᩧ-'F$ʼ3ҖUal08@3xIǞg| =PxT\ ,"z2I0Z .wȤvH) }l@}m0aQHE0m0WV0JVBZ9 pu\̂ C2ש)WHq}0׼dRpsY{$(@}0c(y@"2Ս,"OLQ@~9eMMF렛f|&!42fYp[>z/sЭ2?G{aI? 戆CC.lKxuk?G 4,Zѧ>`h.=f/ͧy|x؅G~Z(yS?xR"+knFYOLQEe=άbejDdQDQUDߗ ϸɌv*SB,x!4KK)KZ\`qLvXfYYц̘23Ȍh/agMҴ%nRN-B+wF_3)) jLvH*'s#| B,pgdޱKϳDVےj=Yqm9]d?O0MZ\]n3Ghf"}˖Q7iȮJ΋ȃ&tŞ {8A)`˂2 3!dHkX]n^Kd2!`zo[nTkϞ1V=wCv T{UzVٓ/>%S9RvB4!aN[9[>1ĝ/#NK:`.PۗC Q˘kO˘!ሔ膛adv^sx(G :2\5 !.Z;v+mZQ~*Y{TqQ̈e9ߍXTMJz%OB(RPm9ykP{A^R}Iƈa(M'Ju10WZqثl*ĢWhT|_b G90g/2qx7oMp)o8, ZCS- ar[Qn!4Y"=!"8B\YnjQ]/ꉎbqЍ|6,9ZK,oG姫Vt-ljW$}܃FSW@@&C3߮wr7bTvuq8N"Db1!CtJ1P" Cp wu}곡EOCCDΊc=!DJV`pW坳D FTG |c\Kwg& T H s\B숈a)k. aélP固B N0??vYkOnMT; IN;CZZu+3V0 u|Xk +͉1L=\!4YBsH&@e{G'iy7BTs3mza+|*QqE?3C3^ \plϵwHQV'&6``6B9jZI֫2CJXJ%&Jb DR$:VMDT3)r̰Lk4 D+T3ёO w#}jH=VV&"Oh}JSQ-p"NT)!xEDDhE hڝ%QഀUK}K)U԰$ tS`+ĘRm;p8O+v:Qt4S%d0*8&-ӁY©ϹE[gy)$CB=+PJu&8);Lq]Zh 3JWe5)AA"p*B o+L"5UY?c[&U FE[aTSXFr~RZ[O4*Gl9~ tƝO ,ے;}9E>S߽y;A;$,{awo$Ep^'9Aј+W'훢5c;.п˻BY>M._>rU2] xBqOXqG+xq\>\]^9F7Ro]&/FnOGK0~*sQe f!33Ȅj^.#b6NoXe$sd/QS]_<<#i37xrc]<3 /*[^w;xkf. '6[Sj0$Z{.RDc);S0FHGGE.fanKaB&I6c8Kל:acJ#A*xd@+L!R8pLƀG+دQ4)?ݚ6e2$8z쩻5bzL ~w9} lfbÅ|X8V5'sNj=EWRUuUvj+qN_Μn-f~$z[ IH/|W=O@M-_X`g`SwD"XtH˅Җi#VA( qǻ\A`jvS{tA@JV18J`2 q-v9R{n-{M՘K[[.CGc7FSk Sf"Bx!7R`![8ا֐`6`r 0WtuxP,.1.g#%gwd]\^f( hÇ/AU,^y\<^VuzH pe1|%YT!?ɬ`lY!PxԧUwOӸ-1D.xhZv@6RdO%Uya*IHWy?9iv^IZfoEdP)"PǵUWbcJ4m,$3$vմv,y ;ĵq2 8(å,PV'A{-IskV3FzoWyÌF"Fz7Ax")T.,ZèDYa*^ taD0oG|'N%s|~ ^Fp7e2oRf|{ rmLިA[w_o~zޮ$X'.N{NTa<(BzjO&DRqi밨陱m:RXU㜝yώAJNZݯAHtMXSm=jjՋ\2 $2ɉ|ՈLvd:!2֯6V\YtlVw%:P9(K<)бucYGmίphe-?"z[x;mַ>!898).5ƥժnaЭGOJ!dkp).56:Bb,}(K|9RŔFGeWů4ϞzJ"M m^t4mu jRG#8}yerNX"r %:jCXqV)ܧݸ(B??εDc'<2/KT꠵$!rݑĒ= e? {D}.Xql֔1 9E)AH-N4׊[#g׃@rɰocH&cS{;ݬEaE˩T0n)s~*M M8y*R4 &89A+ɤqH;$H +<`@25EՃ:qgi1մŒ5Ŕ^ĺό;?v&2]v/Tz;x 'rIQ穭' 4cHzP@B@jĹ.'|. {m%|`T ,]4 ,~;_[Wx1xWCg6;0_3zu5:Gp\ğ&>mM9:Pyy t4_ycxtMy7O/gc.0Q;eS->-r{([h֗dR))e?/kmFEKb oŋ,Ξ3A&f=ȗdfSlIVflw0nWbUU5eMrc,#HkmY) J֌͊/4ʮOWv,hv*k2E9,0cq4ßc_-s$QEGoaA%yX_#QWev(tk~?%LTQű0ڈMİ:N+hJwE7K?o|?Ml:=))ND& Z[PY?H $Sڱ6Epeko|ǀpYm$Ӓ%2:sd3M \9~ !REr(w*FaGW}dzSݔol:^kwt_ cˬdT~/lkint+͖yƩ _amqV2* %#[=^](%LPF@PsM9)PD窺+XchkV$rf%epg3!\b{ pA*Za p^0$b4t=ԨYSWIF-OP4B!Rs<{֞x|}JU̩ՂcN6jqA rO+/<[wvXh`'c'\DrW'p0|N()& `\<3F?b'UAS*?>U%%tj?3s2(WzY2j8N2N@$׋y_W~A^U|;O\m 470(ɴˬ`<"Th ¥+(1W?%S_K_,~ڍձVS7`; 5V^7` lfWk*:u0 2*;NČpEk@sCbAfR3A+ --٠'l&y=ˋN99 *7D}~ui#=>\;g}'ʥ4+7 YMJT2?/`CZKv3Y jI9;-# '󮿣# d:G#m &ZR (gQ:.?3:U94}9??P~ZۇSrYlg|):Y3|1Ő:5f4rvSL+>1:O HMKdԪXZA |%B՞@NLSkt U'i 1svmN?!WXUbxG}[xq TRC{(MM$#ҤVBHv 45k`hsO)k!zj*@oP+jjoD "M'DQ-mH :TXb܂sP) 2TY*/R|cCQm̬Ή'Æʠ%diRý̍!(wR: ]P];q7D6B\rWsTuZ2E!ACqrݔ~ȶ Iѫ(d8Fd#+/W^4h^UEKfQ% N"SThVX30+>(*}7KV\NҼztB.qł~zHu zd7r)Ec#f3^I}4LC4Q)fXBJMkmw{u%i:u126d:VcC8m(2ې9QB.ma.IEn JZBHi# K+*:^߹ARظetOLՈ "L,i63i4+\P"yeA-%,FfcQm9beQ)LJ`ҰԊ[Z1I5nI®jihj) s%xh^EK!zc80$Zys(6j=̻Os8_`"~f?}thksƩndPq3@Bp'(r*pj(v/4wP̂L Ԭǻ)eLpR7/.4gZf2Z3&?y6џy^X̐A>پڈOB"1Y`?I ΖNfͩl;kP3ShyvPN5 壟z:R.p5v\NFC9^58 IFNl\^\\' ]vx^@6DE҉)L5ナxsAE-TɫG;h<^ C|kpq&FW ME$Y';2-U S%bvgTcVNUJ撑[A5TD8X$VsT_S fCC4DX[Yv"ᙿvq6~ohAbB X|mG~4mlkXV~8nR&dz -щ}FvL/x-,dXŔJu#0IT#g񎭺8Ew#`av^jk63kE& k[nd5yo)J]1H0`+/r9bE #(3Qd|QSm쉜 ' +tt>Фƛ<uD򭸾EQFdRA0甶& *VOa}O]Lƣ-у:mn7;z݈E+Wr>VvA/x*aeZeNF>u6 4B ).M xB'OR(2Ua1.ESUPIc yh2[Ǧ)rP0qz(O(jDhȒL1?R1͋iu.>i5,`̈tssgi3U /|٥*]ao1L_ :ݧ7 rؗѢ[G-l:ֵoK6}g#%?Y:AFcvmc.}b4%*@OP rW ՗RV>(ب.atQ_qNP 61|{Pq{w9{{PAg )r1P}ͺ?y]Ԛ ͉(ӟl<}'oMfo̡Ůfc%#3h mNA'MJ*GѺaIpm58&rogY]`-ޟǮ #/)-)(aVQbY2dYiQ|琎7wORV侯[`9Ml9`!+Cl^I= w/%Ʒ7S/ľhˡN.4= "')Am8r;pp9+y V[UYfVy.d0qy[Oxp&JxMxvpV)'yMy bK(a-Tv6|WX=p%i|_רZRؗnǧ&(^SqmKo"qN\эuO/q2o_/Δҕ?SgJW3v|s̙aRK gH:*gR"U*/HE tχTX=}`wEྺڕwӏnMåZ UjBL8F5 /+]mokWm*Ȉՙ)&u2:e")uɤScHQQ0#Rγ`A}P >pAVꏘ/70fN Xd^9y[FUr޽mC01fh#C ubDʁMklC-?4޲J9.L")) 5 (@rA<:sjJ;BLry=U/AiRTZDEHSf&c$caDӆz_EWs>Ἷ3 tPW 7fW[svp!eTY%)8׎PKQj[QHm,}R\ fL rQp ٯWkӧF|[Y)QoPB\0OR]I9;ԅ_nn`Lo;,m*n53ˋ3&!oFdcoxxwn3Co~Al r1 s}6wWcS[ĝsT`< ٻ6W tTUwq!9brѯrPכ`{)QRL3UOUutuӾm*ķUoBӬ^LAÁ6'"xIWˬQD&gEs=mӆ|M;l:۹o@k)o(HIX8r:]0S6"9&:k/H#l J5n r5W|]$괜zէчX(k|_&v gh(殑^WZ )lg~B"2VflG8s^Y`C>A*gbt8 8gJ)gWwA|(wZ4 j;qєC ^Ҽg[< *eF1F gR2DI@ɢ8(!.8.ZB= 4Z~0A!nؚ7`I rXz3ZJϥSUWg9uPLhDǪh_(k쓷lcMu|vHIOЊ}OzrH#+xսd S⎨d΃Id^sAd d{7mǖb8w({6$/4 )&Rq ˜):O]4+ N8],[|,.^k)LłSDRp{ ^GF%i6.Ӹ\tͲpWo7*]"W/и}@'$>"b4WZcAwS:&WHCdv-S&BWlʂp>Svl4 (4]VAmN˨<$(@PxA !u© A5d0tԞSΥc[WMa92R,$ard0,G[VKR88a}uX\go"Ț֚RJ[æAWQy!Mtr\,Y`p^ENf􋻗CŢ&۶; nk'|DdDnfyZ'gΑYy VR؆b߭k_XX~6TbPz JMZ}Vc4T=yx6?W@бl}+ulS;TFJ} =$cL hA yH#r:tCzHy YJ9kTL^ @E@e"[sx/8n[n +AZ!?U8 `+ ȁhFe_AjR&!"aj 9 Be]x% Dڡ2.>3qO7>j%8ԎY+BF)V9򷗻RȬ?[FJm'ws0t:8 f C X]S8C"/)љ$6`HR`qB:,FیAl n `՛5)R&J`@W-ɀ:JbCs6̉16W۩|nhhx \rrI&ϕ}Wh4fV)S}' ?8\""`;~-k(4?'4/%r`G*HvqBTRp\).$9^u(9 ug=XdHJgJ6CQ7 X431KDYIAD<{r@fqV,;'D%P*00 }3C1pFQҚŞY:!u_#ͱ.>u$i {m3s -w1anۿ]`Z@h]{aܹ86/v=$O<ǩ9/ڢu_v3%  g8- xCH1&u˂"PWlF?D.h3^q9Ig{Rj#n d( %>"q@ bV9CdR.)#LJd @Em Z~&x""+%pTH龉zht$Qg$fsΣ"R]4E0ΐSCPZW+baIp}%VoZ}͆}X4*5n=Kl=}v35:+0;m{]{9܏~}0-9XNMOVde7_9y ڳRa{wzyuugEOJ"\_}z/|/\]Ts7,>|y5+Kj܁ƣ(Br"JxlT{ 3mrB4 s6r& wa^`oRfq/FZיx!mA|HGȴ8@}6`pQBkq4h`1B:זK <.~9\ۓ2ʋw㍐ndl8?xv-?瓛'4ş]7H[ZPɥ$J*%ԑߚ3*M.ueQ~j.HH{s\W)M#rM'眆orCsRgLy `@&GDHFuhJo;CzBjf=ю>;,-[M{[o>B^94Z1^:1xT#rvx%Tβ?gӻxr~sOjXճx=g?|UnMݻ3-GDe=t$ظO2x}FS7 (]~wF/ QJ6 ~9![> !*IǷ3ܣh}܊{s9- m.|Ult }e8!nawogRLƾ%jv3L'űzȯɽ3LzVRVy 6[v{ԉ_ ݷ~ ~U͡ 7nDK i]C3n]ZZw[\P ;ss/<jlٲE5 6.†FP)&RY?L'P9Z' eSX;1U Hc;6H͜"XT |]>:T"kL(2X[G#{^iFMOl7ͩQL!@2 <(-N2mTSR=Io(7G@VR۠tkJ a4KWV (=V0' jM$!p qّ\G8ݻo~/'|b h|'bBەrcCPY2 ^'7i$mD(Q$DԤk_QXUg+L6!^cuzƏo?sƟ_"^]q9飩?N9|U%~97<}waY?< D9Jt5UgcP`P&gbXP\,{nmf:R:cxRA=[Jgb2H(BºH2;82\$:G ~j&bkW*aTndh.Gfc&'642j9)}v 2Ƙ6aUv¹Uf3A9sI|ԑLq_:Bi.ky<}Qj7wzg4y;.j >9{/}mW8;Ҭާuy*-?-‡2I5V䈜'o?=Yƞz!$]&X\ Jnw%x:Oz[jr t:uv'K[VNg1+/ :孿_jofkL#pй"ȡ1r!+y>,(ڤP#bBD+p͛^^^,.ݱC_Ͳp yuݻy66Z}^p|g VC5nft @[4Ucc9(iJC&<4q'R4>% ykHEo' :_XN: !X*'dѬ~|+A."MBҘѓ?ȭ/d$h$˕5{b'Z;b`Zy_|)pNc?@УSps>lK-38X8łz\) ob]jY<(?GZK $xioWqoᨽ c2W`6DbV Zp[oBv| 18BZ!S\@왉˼\ f8ˡ#Cp2B?*Tژ@n,6l5T{NJC%цQ P:Ye8ZF BdW7\ڀ4,df0 (iˀ7{ɚH+ ~*ǩ&VR0a[hUu*ˋ2qݠ>`[J-f >e#a2oJv*4oJv뷟FmB-wHkA3B$&RFP|#ophop96jA4)STw6Ao\_V߫GZ_X0eXi^_Ӊ|@P&oV7EvSߚD6ҞYluj!TFaX^_GZ^նXUy)mZk$H.x$}qoWsR{5CጪqLO*V VM-"xX~ShhE6S4ĸi\`PifK$zqfuQ&+9avf̌0r¯-d1{2BfȌ_aUEUMbWyZR2R oz4Og_dir3oyS][ ERTY eXhFd:PB3YVcc'mKIž2ڿmVZ"ES 4g{" Py9M)J `{a/H Vvȭ'Œqݣgl%@5tE\I݉:s&Jx2r{/(T\Ɋq9Y1cK1Jri!)|EKJ[( _]H[NW膁3u+NkžB )VZx{NXENTXzFnn*Բjx|M2D:t7c‡0v (sjY0T9H+50Wo90٨Mձ't{*Kv9ƬuT#@;!̛ir Mmc3\rvrx]!N#]p&ʻ"6VgsGQ%}Ý7zPGn"e`^:uvn)w0{:W մbDFBiFtF+ %}:MS <@p`G!\!2U>'YX5qEH[ۅ#j4T:xdؒCP353Kricvy+nh-HDE*&#%KkJJX_ S4J k; ~$K\`Trxgj=P!QRI)n8S}.FZt{ 6Z8,LADiA̖׋Xj*GA 7j}džB,5 Tg.K}@EfKWĕė @qM}~}} ] Zt y,K ꢖnm(sZ_QFDe YG~60tJaʾ4w5xB ]ۛm\7Y`Q)Rsb*Έgm$$ƓSW&^WTՓɕٕ4;cX%.AVlܸ5<}o]jm}1J-sGGRB5SSiueH $ 7.Zͧ1^l[VlFLᐜ-jMk)#ޣgf}#[eUETSCX_Vq 'cW7w5bpI#IK7Yfqs2~dk(uX'}=͛eqr*"GaCJP8-t6/*|kC0㪜аkbU_nP\qqH'e\JH p9\5y<!ຠL7R4w7I]˭Q>b*[37h[nUQCseg!gnŦnE3^"WwU,wB7*2pp L^Ne9^ƚsYu捭TljYБ0Jym'cl>+ܹbpb++6łj <Ųdei1~*F-^7} 6O]?8^(4WmRzu6jOzm0ٷ; ݺL91Sm2iWbr>)At`"LNȱnv`+ "N3Qb/t yY.h"1v__|HQ|EŕzF`RU%"{doQ`q2d64Nzr^1_*"w%e:[#<=+W*Y`0>~wݟղj|/ +ґ8C*r)!_C`H5{E] <p*PTH;,)i:SQXKa q9e, l*vD%* : !LX^Ğ1~<ƋM4uu_B-X)!lC˨T;lupr])V)L iTy-}ݻ5b 8.I2f{y-`s1 -:?h5'y$%0BSs.v5>ۘwBF/%wD#~cR*m})QTbxVc]m)ѡ奮h٭oE,[3"'z外K-3M((f#r$224Kq䑲$p6j#B)\5_my6RRᩄᩍOZhm<^J=!Ȃɐ kUUl6KF IʗmQٗ金LqiT/!c_+H8A8}tr``j5DN\kpG9){\ @A/Sg<zUy+iRB5dz[$Hl{ PH'V{k-ӹb?x{rQNM-#eh /U2 W_ׂ x SР[cW0ZJsB[E3٫mw5/(#\H@7jj=Z `+&LQw­J?& o2G1*K-h0|Yl}uKgx]t]0ry7bI@C2 v^?.!SF: Z߷ثH L4jXl͇?Kn=و#>2Q6t f[6Mӕ u~qུFD#UĒgѮvcIiy?T7.1$ H\bjXHX1~zVH {48P뎄Ԙa{muzT&U*I#Zrr:I˿[9Fэe0?mm֨xd3/ʘ Lkjy{֚kg'NtB?F%͆h4VDs }G`=(yc{ŌVؐ x3|"\0/]Bokzb7s2[NYɒ!2CzBW2qhdaόH.;V1EjWU*wBk? d֜('jUj4uq0[0jg06&Ե @59 kDTz73ggLkg)2éZ!kiT0XY"Z`t]~>:-5Fn!nMU]lChKHo5\pRQi/BJh)p+$cs>]csZ#%T*^EG+V' RzJ!ׅORR/>oi̟Β\2e pNbadx&-zs]or PLI8 HqyJst<; 6Vmm Ԭg:b刉IJqJ s:Rs(H'`W©ž]نJyve_4.YCLKs9AirÎ6 2g\ c赡8xLVQƂJ)4! /Dr-gGܓ>Dx=g"KBRQI/ZS^I?|/"ƼBgJ(~T#s,WؑΠ븻aRHG%ACpViD7`Lэ1\ژ!NNJVJZ#J9IaD̨&J*\#6Eyǐ A%B.ōqنAΒ/.S0WZ,ɀOܒq-ǵ޼>Ye1i}:5gjxh.Z+,8EBਤmFsRM H@X :(4_5|W\[fY74GRF펎22 c ?9 WL-s_m^"*?}p`[~K:q gfp>Sl{s}{Z?<4\ܳv]{cs}c{r={inmlLwްv͟~mrO4a7ͽ c:lM?OAOmx9. G8?52q?oil96opWڑHc@7~,]io r@ʌ[y5fq% fg?_JCg44wDchA£S5ybQ>bo7J9\<m8ك7Mbz9Ak<ͬzpK^L>cQW5r?8˟k.扒G9^4_v_OzuLw#t~Њřn r(f8w=\ \̓ej~1WHt쬜nxك^URm;sݐL>]~uc{{>|$;<Br91Ӆz1{t|hn|۝𢻃g]eis;sS{ Go~L?ÓgQnECgcaxg2蟞\w`g:'p]L(~r>]7{[s,;睞jL>uȵ~nҧ16ֺaƃ:8ъ!iK8>>Od,.c ;_O1s\wdlG)fsz n~#|/]kB?^1[Ձ.("뚐PbRϯP;\/9Ҋ7OՑPDaH.Wʅ|Gk MxYy *( 8+kTUV)(O$6pj3t۲ۈ ˹ki466F"Џ%vDF EA"."wRšplC"Pe_{hո l~ei> H;)a’T$:$WQ sҼƠ6uD= ZQ#juD.<16YYY,5@>-%;Yd8V2%g3=  l2UGVxT;oi-pޮuD6Z`dόW0VL/d.4"cdVnu#_GVnVn:my㦈2Y 2c\e Y/Gq8L FDysWud<#|0XGAH!i k0`5` k0:0DF,X`Agv~/A!" I#2-Op,t5UWPwn%$f*aZLJA̪iv:NY)t{6\ ̄0) x9ä,PlD3iD1`@Q',P$%_^rL(0[%|jPB)[dQ;)0KLpq!9dG͹ZZ*ۨҭVyd`gvi`'1a~ bǦ~8w ghތ:<>_<2A.V1'E9Jh,+!Wb%XIF*@s7>%!1RGKO7Z2u:󧞺6"muǤv])(.6 [Xkrw?ɿ:ܥ;R-7(%KEL>mtLj $"\@h 9,YO^(P)?l%M"0wzyK,e #,O)R5YGmP1pwqbL 53|\j:L`̮g@RQD>xgXӹ~@VQ3w"5s ֛ʘ86""'B5*R3HJ X iIӼ@iÕȱ}QkEJYpf"WHewQ".nh o^`1P f'd(n:/Oj; .DV!B2̉OSvJ)Q= ъe/&B*PPdHKg^c`<'a>@H& #2:&^227AUMWrJ\/^j^puвT-KuвT-VV 0`jSI]v |€XA6`((F1`p/!j+=oPXؘx'U*"8t0ǐ{#0uEv4̺/}|FE>ve"I0+Lhz0IQE`a{ (P>8-C`8[z{ӞROxyKZ${A)[*RʪQE杛'.d z v)"qz0e$K#;DMcyᵔ pA(Z>wQ+*xRz2 tsRShqqQjx )y3ВpZ L yE(""J򁔒?`n Γڵ~UxpbUp6{wR @&JHIQUAf_ FG2 `-Rk Q(`+ mAt@hck06&T*cбVZT3pe+) gZ],Md :*g)@Bp !#lrR+)qD D)`$bx݄+E]mQkzk/c쀾L71Byf1&TH`BI)D FVZ(1\ki#FwW4ú83Ѳ fG(W'flȄ11vwuy~ԩ*.l2$Ary&Bȗ"Zv),b[}i--̱⺠B C /$Z=BX$@3=phF}$8C)# h} +DZORe2K/+nƢ;8WV2ɀ(2s`-J͕& ~ q].-cPV>uJb3AkVP\C H-Q \ZBR)M&d$23%E,K5ڊ9i+6)DBn+R/H4 ,= :q*=`TkTo ot\q?& Sy܅2S|U̻(ѻsM5d׺0jZUzn&6BB4 +UEYlQ* t^+\C. yUI'`zuafՌBʇի+ \6K(B(yGzuewWW}Zt" Z¾Z;4οZ \ zLz]3j1S=9s5j=9qrSp&'<~"L\A0BbA0ldSݲ 5,t8`brAFkuMF/Ƿ}Zqى;_?T]ӟ8דy4[?/.R (xeﺒ X8NYCBVN^'-$I1׉m$5zac݈>NԜUmR)Dz'k1S>׺<1Qv3z.sl# tAM%WLѻr%S;q?N%btJ)H:C>'ְ1>QNKgZiD*6G5.;9HZj0DJTkk3Y'&P aFINL>wHL%\<$h8QV kTܻ}zIF3 fbuf[g]g&浮3OG߫PLEbڼ#GΡ, <SFnR,0O1qbù#Z·}N7r]3 Zl]Ny uόR\QeGwٳ(j4,[:),:;`-}ņOCͅ9MRAw.F $uqe0P;4k-8};sh)G<}'k]ceg =7$ܚ'@ޫLiT+,JyJ>)a|iݕ|35%**Yd@yyuκQ{GF!s)@`//.91 Y^ |[$}>(ٰ\}uf7DaNrub$.Oúx#(nB(L0ѺR3ӳeg3ZY҇Tnizb{JdG-lT\h> i6^xߞ~c0hV"d '8LsrX{>ݍMC]$G]ƯGʫvӃ5{?&Lpu?>Slݭ7nuNbאf٨?zë֫7o[gطwFO[hq֏?{3 ҭ/zrO}Mi?Anݯ6oa/nNiq~.a qh6*heg׍tUwre45]?w^nxzk2K!Սyq,.矆^r瓟np98ﺝI4mT︓O >>h*?wꫧF {rrF>CwOi n"``'Uyϕ17'pzIч|>PUT<'J?n47!$ZRLo_G$ݝ^Oڮ*u^aI>bd|ec'۽x*wZqa׉5}} }?@?tt]S{QQ¥;AOJ=&pN=Ox,RZE<b,j=xҹPbY'CbB[m(>DQ9x<%1Nx]i$ )D4'")e;2,$qs8IFG\ӨKM Qd0ZȖ&'7\,BŹ: v0> &MpQ\;_a5q1r UR/X4Uh#kJ6^ @ǎ_<96d$;}_g=6Rkeͼkwnpmok Z^_79JFoĄnMahnaM7 U{萸iJknKZ2DL.$v) -!,|3Q43^]AKTǹ#/EYrmIDKoZƘ ˜]1[KK*݆jۇu> AX|:1}XePAByOe[m<hMlW*<H x2,[du0E8omI@JLFEĠ95qWg%sxot J4Uf߇XR+9!.æ g2oeF p?ﻠFh6z @+  *ڝgOy2*1OC*284F8\` O.FʤI(}&9iKK2J)l n*n!Ts"/QZC(<C7ZNJR$d1H $s0)=X tCQLrmZXK JRrO~bl?A55诸O``#W'3R;c*9;t.e% DϊV04XKi;QHPL2KWTAbtщأkXg x mmTh+.a3S() 4$#s%g|$BܼY _g)/cXޞol1r/>,, N GŗGaJjlHM͕_ВSC a`*s?<‰)@8*,l`n[DVS}DQ7Z,B%E'fOIZ w@,7@Λ@x9$zƔ$SJRc|7J .ND /;Q8U*K|5~˷E JEla+-м)-qY{`ɼOB%W /m R-eQ02,fCA"^xT1gͳܓ1b F7d)YNltвƮJ,l-YT9p7sKk=ےY$!5,XZ4:y<[ے]PwVLi–\/ Kv˗V/_\|8y97'Y ?pwG-U+&ocoK (1 :aNeF2cAbMd ~VYY8f!I"|ͫ'_/iy9]c!i% {TdM${9[S Im#^cu&V: '#@}ŤCβnQa*Q(i2iQ+$F% XuI(f״MFf:1@}:EEm%3t6EA( B:-p+S 7;8>-+>cM^"xYhn@QCZ+-$ 5{̌(D3f3 a H BtGB+ n]32n`-͏URÓb=*}r7p tI7,{+ 9dAXH$VdEiW6ܭz TV6JɹH)8uI-{TwL|R? U,<Сf59Td .Ry\#̭|`BPg%iMcfق/FX}6~.e~D>I'^6,`;Zܯw:j:[ gWipB`%+y$)(^OLN,f%y##-79[ Eݭ kZ9|z_10!TΞHSL{,fBJZX πKN2ƩiG5li#k}DEU|((ϓ=DM$,S".:h6Za jMR6%^Dl1.@xe5rIh&2i 1akia;zjQnA9xETDP֫]Ai >i^bd29 ƨ#\E Z ̡2N)P@ Ab ڀf J Kl~u2h-] #IERPHP[X1;ld$!wI l?&{7D’YzcWF*1NcNH+݋oP/q6{is?DYErH4^$02FBj[!Zxु ! 1I-F! YÆj3ؤ9-V2?YZrm7oпB:)_?F44g.q v E=L\확ݜΓp"UޚDfz ،e 3B.mj5`T?6a& +xZەkfY)Eqf /MGD4-p 9܊&ǢG |3d c; 9NQnA7,#cp׃C a⑔+NCg>03N@dTFA[q) \7AJª~-c%xAk`2 K;XB$@1jtv8%s+jݴK%zlsIܴh\%`.;^J=5K+rP4T5QL'j uDbpR5&bi^󘤪>j`fKIi,)ۭudR /,2  ?MR )747ܭTJtJ qȋ7?Q<`'! @w3_mh$h#I=fǢ)qc*7J\#[L"?(AN@ *!l1x!DʸTk۴*i8gMJMp(VWH$Dln@X*ME 5J6֛Uxxj1sRR;FGZ !PZ q D5`.-͌%S#+ՒJ ѫjP)Q4`7VV=0`7QKC"* J$ \,Qe,TJKP+6:ӑRd:2۪R y$P+o[PM0SE :o@pXNPO&D*!AV[N),P*YN GW(TQ)0Kd!O؛LY0uK,㦔3u޶[|434J8xC\]ԓγ}`[Ԕu@-<`b:eԕUٗ)-n츺[n:۞әx"[Gh݈1u#$!&.-VӷsdZ?C Z:VLnA\UG+w84^+MIs(UkKj &NཔdM4jAiw*t>tTkފ~;%!"R!:\dF$LYJ t&FNtL>2-[66C0kZVHm5c)$T,+H?. ;COSdB<GLnfiVt<^bw)tW|Q[:ˏg?h'G7OB%ׅ?(O.e' W&QO=QO .OlOv*3v jٱŴ.=F3H-pfM[>ăŏ91SLxԠ7(!An#RriD&߸`f :MfKZ9-wZS׊8N[5 EnZ<̊[)oi߹3!^h^a:^z; >=O+A)ּ:DE~J4-PX|e[.\ ܊ #"Baܿ?Z,rm)zo;J?jqn;}W\F|U7Ue+zZ&0,![oxgK:a,S JjET!g/YJm[\f*Wꔦ[khw7FqƆn} 3EX&S 0'=6@hB79U[u]Տ2 #NOZr9\L_Rk S,L `w:-Iy.BEQH5(=BǶkԯ|W(fx7^6Ɂw|{ZF;ٷWq}~G)y]0ANym{o.>sATf^ƀpcn?ۜ{m#kW[k6S6&T· mҰ5~€0)r_9'VQJ.@:!Wi`dr2ju'x9ŝs;wپ !R˨4'4RqI fJd!5%KU*/juBۗXuqN͝;E7w\1X&B EU*j7"Q/$d e!RT/jwUF˗m@?Tԉ`\C~2v։j'RĻͨgpoוe^8=扚ci2wü7om;2zՑ!#Ͷ`g[-t1.9,}Vc)s5[Y(%?{W6e*/[}TTukz>=2) m%GK ʎMYEJCl%6yy{qυ2/8&$gнF`D}5J /fFEel^\M < sCxO[lO"ڜܵp͟s»R±77jWs 4{@2mha, H({~XvQcʵ3N 0P;iib c=^X^@PdT!` Y~"6\wuoE#d9/..D Go)Eo>qqE luy-pJ/+umHY?@GHӼ@b9?O6(ߘ3LUK# ȁr; e(/sS,dyvv%YlڻuZ;Y,Ac ~#ćZ}5[_n<[U5]51΋ {}QS)gz)lQ ݻx\ˬ`/`%X8](Рz-<iEaRf6TQ1F93Rſ/vV[\q:BT5xqR1?^^Cd=J!G%LH#^h8/i b:#*l-y-Vv3l,Rjs֗+k+ټXPu+ 7nC<ۯ7_VnsX<_y<ugb[}Cvh6^d%|cQ|B ^63#Im#ޜ׎_͗OqSv:~+~3k~(Cu'Ï>ǯO$7w&godY_fwwUoO.!RQ=KS7~[:}ẂdRY~,.>i*hl rd+- kesp^tUÓbݿ3"mfk񮏁!X.7-{D=v]uQ^;˩[J{ DlEYn R1l@JJ#K[Y $[۫߿\N&+_[tTF6uvLDW4Q2Uh_ܠbnc }6jMt=p_IsHp?JZFJff@\2~A0,ؼ'߰eM"`Wxq'0\%b30\ ]n:jN 1=éPfp7 oڔïx4Cň%E#yye=X( x4 k͈ CRrHTub@Տ &% *ԙD7*Ox; 5mvQwFa9.&'ȧJQ!_<L%}&F:2gF3^`Ň{>LfuCZhh/qcn)KBuɘ J ڙnj\Rj䅕Z2˭\Wu}00L۞Ô?5[(3|҈ux8Z3H2,TiChw7esA3oD NrT(H4s6Xm `qMzd 6*6 =R6j*0}Q"r1|1[-;ةRQu6), ܠLNtm◝>ϗi !99GD2#FLJy>wn 9V(կiA pK n4* +U ':"f"L̓ūUty k}x^;L4]}\\:{"D<yOyy{qFS^ I"OUnCZr-wVBF!֩)\:#,]NAu'v(S)rve Ibf=)@ܴ߃7'"byz 0DD>FcZ)6K)-A -=фVZTib%ƤPܳ-p.hLA39QZ"p.MS3"i)apqL_veE`5թ$xY;+{}ewf?ȅOg{lS2Y DڗB,_XpʏV)oa&TO(V`,D`ӄtn~g5ˀexYiuHQ#s%,s~7vܥ9D ˑX "4S{],+Afʑ0 ]EʡXaN(; -<{`^a" /Z|‰)n=dI>xJ~-m)Ai%m|MQL c7 R c9^{%[~e/;U,A-bm` KkHqLGONi@pŔCU>#( Jr6 uZg}p(6ȺsK5TLH9,5Y*f %$-)cp &Mc ,?,$-RaJ Ilq~ !ը{wy$ebW K+g7~BIn3xVZRd2g$T74t0󊜥̆;8MjplF\?$e=W`A6wB1 Mo_[$}fA:| ;9o(%ǃ>=dx;'hPȧ$2UZY1N:ОX)&cy4~R<Ь &[3]nx|-jcXdm`21a+Ke77bQS]|ݚZoRqr aZI SeV Vt2!8qZ[]kR:3' d)P" ƜM̝oW,ׯW\:hv>7Lmdbo|ۅ,:`N lغ/:חWB{XuzoZJuNh'Zd-TXsJ{?# +JY"N1AA'UtTwwAߏ4W(vnuJc"z/:y~ՅOA{~l§@>nk)t4\"a1lx()'z^/ê~ձ3h5 a?ñR}U/LŭLacqIH!4@[AqiA!!~ZI!/tTSqvJ8 ~j iPLQ u!?O" 4r\ PXSTE=mS,\F  ue˭3\so_nicL}1*v27KTέRf .AH0iZh 6z{1T 0F68ZK⬿/)pUT/g3R8JΛ8q0)نPD0$IPN!9ibk!6mqQ!EbylXX$঩|1[uUc-Lk^y|"0}X"{]wj.кߡb͖X*NzS kuBx)l݂69dF9saɔDʅeV`c{maHŇQ41'xxg#V kT/V#+ \PEpEݒXUwb ZNBWhG=a$#TQK;;1>bOH/%[3Qh-F02Fq{9bi|MUx)Vy WHY4[Bqpa= OWHXKʝ@b)-SL2g2dR s|8 ~M4`詯Qda!)m6tLrۈhRB q .OkH?:ҧB0p:λbHⴴGь4e,쑒[H3q鴺DRF:ml %sw+NT]iEc2xwyt5ܜ cw<2){%0*YkqŰ͆)B OW"?'kfGXXf_Mf5ׇx~2*Ut9GUQ_>2'R'~2>YW+-d9+K\`8`[0Zؒ0͘?LuYZd6{<ע' Q9U mbF'Wfup'Οݻn'nmUQj2o~qPDI]UK-C.!k8]˚5^\ }(l0Z|ycW$f s{|2WlivGV1}jF!cu"y5/H~{;/QSmhI]oƲW"1}?$ NN$m.ڢҵLɒ5!)۴"K-m&Erg۝v:% thEû7?}?\ٕ@]\8e }9({Gã[$~IU'W_w@yU嵻tᗲH #<-X:a(8Ʌl9_Wa)0 st PKvFU]{Nq)D8biT#ݎ(#Ed_;Y׺Ⱦ6m~ƛTŒןL/j8x/Ir8f#@e"]L*6sGitH(h%ڠ斈InwƷax'>@e%bqw&B+ހ΂Ry43&IE4y6ɨb쇾b( B)0;JGO)#!7t6MgRyރαg2B@޷WRgS2=[8(Uq]4j%:b $a>i~X{FN4oP ܭ@0w?Z*# \2؉ZԨI"rS ('XJ87fC*`DnmiD%oDeV|2 9*K sKku{O)\b[ Ͱnϱ2 uQ[U:0zw)B7@|2eI Ȏ9mAD\4z5]l)i%!M܁Dy5*S.Tjx6lRB~F?47K!ֻ c1HG"70`0JG^x "XF!#"h"b/qnd~+jܹCRHKnm| ֤Uo4ڊ@_Mn -Ѝ2 pn_yXIrt!AR%e?I3kscƜ%!?-Ƭquʼn6*PP"ֲ&wvvhVelxq>]ǽFHn)cd͈cHսOecDZ;RilOXZCГ-C`I]YJKk)Z'4!)kRVJjK,C 05z#4 ]% C #`LP0 S"y+JaUr@&U"#beBj&zd~Tr v-)Y+6y~z|·o;d&R7f,:a%U+nQnDqW2#3M h$CtAthi;K,[  ܬA,!i] /Ŵd{W26T'' ,BoK5]/M(C$k[T+nRcͅ%7#ܽuiɴ/{B-egTt~s6X< 숺6[qҢv`m@ϻhDXsm 5o%Q``#sP)2KL*nw)lQEXR KYX(!0q|V;hBb2[a/!3DN.%y7^Dj8jڴOtq˘sBa.UMy2$\Z^XʠXsKZIU>즃\+mOP$-vCgAkD-}=Lw'wOC)G7l|~OR^brO+^H2HsE.bnvŮ Z1nKB-#D˹FWCɸ`٧yKO_o.^r[6.yIr^֓è]KΤ %g6F6+mJcC;džp۴.Q/IZrl{gmhMgCeZi Wh~~^pچ^y yR]ٓxyr AgkQ?8F`,8 TM'PR*J6R*h*;h.TvbGU6UY\mܬm4Uv6咒J*3_T r'y\Β,xXݫ 0{pRgkr*$rsܒ \HDTk.*k "on#g1ei8O0Xz8}HW[JxEmd)'n$k16Ed=‰>q0jnsq5r'gˋCo#?iݙW \E3H ILz/Pps m Rv7VJ ўsppcyι>IE2PExDMU`6"Xit3a]_,0m)cﬡ&eIKC#iL?d_ ;AAŎBKYW!g/D[H4b kvQ+֩#*,4bGIz0 &dGvb8ysпbk,|\׷Aْ?l"xӭNvIj)Vqj+a2ҕ5K%J ni˜pT"3;][.3鴼5awd3rJq{YB0 B:>iq̹9o|+F;V JumΪX7+]O,$4ߚ2u:ԉcmO(V0ОuvT|L4Pa\4ߧl^ | sUsێb7 D?6ź+:pM{{nY(Ҍn›|sΝ45 Ps _"cOh/ J>LÁ? +7Ct 8>dnr[ ,eqtq^eq8.8m8~/ }?tr9N"S89>ZF)'4ÿL&~Gn2 oT2˴٫ۏ7uH~>N;ǯ `0'yr~ ^g#_L`<q.]e4],r^p/sYEhBNzKK@c 9X݇z{i6~)#xy Թ-5&AaNwM'WI˅3V/s<%^y |x yMr ߿?)j۟O<͝\8F FHq>r\Hz#y>(#߸_59!F"_*U(xk G>RY?Bq*6W|)q8 >tC0凷8>N.GChshlPai^_l܅OË<<&y7>/{&y]|{'X^O/$d= 7ϩ2xr=\/Tk ;=MCL o̠} .W]AZg=dxW *_GW%O܉\Bޯb ^^Mn?n?ɾqZҿ60Axu-齙LЇd_F?OAp&H*%gkn^3KL?_C|ON wY""er+qF tbrҏ5XLz<;Y~c:Ko-)@=~^eJ' S%r= 3:]' ͗t%9dK?/%3L>,MpgX&zY,};7I]k1 9˝4ZݗVdgeq`٪/;#${/_wn;b.8fGaeC~;TA`}%sKKA}> h(J N,c?9,Ƴ|[7+%BQ 0H?|/Gkzpq cP7Hő %KQJl0mmt<Q0ID e\ ~?؂y#3g9BenUg' O"qo2%oJ.D6>:q+T>)^cgHFGRwF/r>iY }"~T秛5_C>7v>ԮRKH1w=o< u2t!PB]{oG*]7Z.^pL?M](!);~3CjjHJf!~B;đmE֠&-g<6KÓMGI 0ajGWUY^:~5+r3B+,4OGDF,EV'LaE?k W RC &0*qI)G)jIkM_' lnV_ S"?pn({xq|R=ri3ȞJ8?9ya'㜰Alft&x(U;p#3 6ȉH4TĖ?5t_B@ 7 1FI+LҫVe~7s)Xk郁Fbc4ͬJ:mNz#mF^RR @n ܊ y3C%+4K K%D#Vp'e(28s O]Yam2ye8s?$/16NUڴFHJN!(4j!𶻲0+C.1tQY{"7ѬթQZ5fſu[6E*auY\#!lM [vvp[^?$/?D_єT6ӯ+Ly=.|qs1}5,yד'˫P;'I+Ʃ,v>JU1H",YFxmU㈡NYT ɨ:Ze$ DNM߄`pB,m+|l7Sq幏hj]]w5 kyÈ\QŘК\ DYLi瓝~S).NŲ|}E.'J0d?OSNӞ x$g=mKI:'ZēSXb# 6& Va,GEJLyOQs~@=- -TF ׶eqJІTM>;~uZB~̖X Pbs 7٢8a|fY0[Z-5`4n$;= xkXS+c,F*lZl\dUG~m]k:clvT< ] ̼i(Պӑ&f^= A\C8Zi=pB=,a.#u( 0EJ&IkDj+"\%wk&y`ILĦ1lQ+ضb}:[M #J@{t׾XT1*ΞZTqs_^p;Цd{G`*{gG_=n>V)҇p,=mֺq>@!vt\v{>_f6=;IoU4'he? -vf<UVW RS ?6~\&h|򅅳Xe)%k~Zi'_r߁FEkld4O [WZf 6$V`\PoNZLJowy>D6")x6B_0A}JKi~r}9P\Q +T֜+N}(J w o]xd8R]ϗ輸N_diQ۷/c}7gJdNq|W) W+;xet=y d@<[m> vzhVJ! bb| ?ϹkM$O; TWQ2E,0)1UkMVWfaw-F6l0-(];[:3l8sݰ12I̟LΨ0h.zjsJapQÍWJ5;nC7Z:LjrJ:8'g!iL(Icr a"dqs{:95QrncQ x(6]"3etT%"s &vmmJ֯\8Te($ʹG_Yl(kÀd־0F=0nǭNK&793i^Ni{9oႋ5∙xA pB v[M GMDKLzH2b|m;o1?ا-l[zv .ZRXc%cSFpo l Ky)o泣mGfGmx+dcQ|Q };v[$@x<]f~-TKt'$SqB,|&bvđ:i B 9 *gaw-ASt wN5OKq8EHU`woGg|L>  nC–:Unَ]?ZA ͧ$6.N v3~A<@+ϕFiQ䳽p2RcZTCv׽2O,) !XUgo\Z\Ul:pO38O"{׏&~TRU9[yЄyubmslvEhx}x- ~=;{/л| ٓc<}4/? Oq7C:_:/~~պr?5* VToӶ3]1 o?zy"U0AfȬކy1WnfU>tVy%]އ[r|VoEf6l3'{tg]o73ƥwv lofty(ƨ)iFajLxΒ&oƚϝ<Ψ*X5岆:5 .V?owՇ#Qhԧצ>E ,e+ d,.(Vyt yg=:Fxo-DiM"ٽ,krߟp~;B:ݗ"/E?ܞv|!0yΗ@S!"k;|us.'Ƹ Uo>;@8K}d(+v%'󗃐Ԉ4v`C:^uql>{xmrYv0p:ys~ZWS#7(ߠ <)XU6:&węBTcޭp I|{Lꖑ{Z^9+Pb8"D8eQL `) fe2.:E8V SZ JGdXuԨ{Dm[#lQZJ s$ O$"6J'`ad%Nƛ g [0IYl[jQ_"Ȁ4&q%G Q(U/a")E/4x0SԌL!]Lrz8%'+Oz'77RDN#SF*FU$$Q~PLtQƬ)f׭ sc& 1,&B4]]Bwޝ<?LLdF}`nQ4R ,mx f汐r(#r@L3!I(iP b%Xo`c |{±]`&^'rf$6}+61yhE=Oy䧄k8L1lWi|PNrz&˦m6F>B)v!)K؈4DmaG RPтt4oMIR 3p$.b`Q"D HDRT`/Jւ3aƥ#eddvŠ *C'ڡJs}VO:ſg'QId7,[Ь6Hüp8x"CClF6ZbBǀ%SC'Q X{8(k≈Vb2n.Ɂ9!g},S؂v3ZJ& hrBLG`aQPlȠa- %X0,L}Mi*'S9 TNʩ2Q+Ȥ:,\{Gς+NP+NPFVӘuq9 #E+2~>s-d0,HŠhgb]U4f&ȭ Bi1A%eTl__9ef\s0kjW%żك%K*.X`>jفfvRdl~Z8!!|1C(r[j-wwv_#:x;_- YТPɾU'mReb@h`cM8b6%Y':z3Q֛>4213L ƧCXa(247nZ&XĔha&`cd9I6/xUbbUAvSP\>tЗج@P\OgMlYi Bw;wVB|#E@ӫep4S0;uyC-pʕU wf>,C0Yj+|Z7O L3^fHw\["pe4Rÿqg$KK&ˆIվ({cðK9fDET j=]u_Wn,;,@ GalZ= 1aܧf>nY bJeJB"IġR ”G -Y *$?Hؼ*47_~3_P6YFD>r.)F_(p<1jRv-8=II2pQt-CZPhPY)o y.74,t|czꍧUe:7ؔXf3} |QA)ۨ2|TTj~X2)~JOyflMXZ*6+_}sW k_U;Nj)Ig8]"\; ;..cS[bg|ҙY[\ݙ2sR3eIqՈ6MJKpln6uA_&R]w5Xk3i`T~ӝϚUӋ0c$ئg^>K𗎏ֶ]u[4xTyE˓MdKx#0 7[%}%tW<&\80Qc'ٸ_/tb84zTdc;CGEIN&ƛ^牥sny5*dyi6oazR,oSՓʒ',:b6_(ʙݕŭy5qZs 4K`;Ai7Љo #f{^]ZF`SG <#?O2S,qKhN"&ie`gX2zCɸYf]Mƀk6r 3-%sqiq,F{3C%]1K"a6îp[oKlKw<pƓZo?tV2j_Aƅ. &XeBi_H;rz? >5?PɅ,**<,7SCI(JT1%@5+ 7 9ڤe/}&6A]nmKOHqEQa[e12M}; və)0q=jr⹒2RA9r^Q>hçUrEYMJSoY`Y`Y`Y`gACK*ݤ`;*qP*@s-0Flnq#*f:?uo pelg k> \sKx,TSvuW汕1nbq6 .6/vg{*XA*}e"rO5OlE;n3+;w !^DKm4D &cbʿ^_H65yʂ7_fgּnAs$o[9;t% xQ3RpsiTlU6|6חlpP"62} %,a1WߙXl^}2R0 Wk+\nUƁᒡ6›&6 yKi)#dNzy~/E&%`W,gT7#wÛKd^@ʮfpk=j:I1Tڙ4$G$4?UϽtƭݳ^?vpÉ?zl==~vxwmw?><=}pϋ}=HZO 0&? otI(U<9q=jP/~.wwUݵ'JnH2շY;WFMA,$r( `% AJTfTGC ".G0HwUP8n/ El$IҭsP5Yn:Ыydo:b+Ucn t;:0uһR7aW?whj}ySlwGqzQݵ>{@L*LnLݸ?tY4ٛ+|noOќ-JOi_ѓdtO^th?[5+Tfz z_G8N<'s[w. Xia}MdHB{YA|Ӌ~^ |ϧ`ח`_>S)=CX$ƱϮs# {(lz'N$yu尫 N+cOm (  '&5!0GGSX٥*1K0[FLf,aAd>%5,S=ɢW`Ψ D EKt.+|6=Sy_[04AY&i#U޼/Y\Ź' a2>þ\8oJt]PùI $&C bbJ,285RqPRY5Ct`:Fn4"ߓ FyD #gӿj2@!ʜ(U&28<V߆2e0݉Nɻ m7ތ.d&ũ-$`XǵBae<`_+35D#ocNfGfbiM)J0K=M_SQ knM UI8T%{^v.trKOSMȾaDzZ0{t~I6uꘔ˳wpv@φ zeLl>n vAmSXEr'~>gc}a0cc ^zM`{=m&=ک ЗVP_Kêw⡭ ͌ky74"o _~}zuE!7\.;[t?t h^g֡;a\Wt<wc q/[.pRqt)QS 5Ъ**@ZUF\0H`Iawop* S0F 1C5ADfsDj~Rx=]z>J=cGؗ7{z^po*^GG4ߨhԔi];w?F1=YYt*FW|{Vr%O.oFzIQQE^qb3 <344.")9Ǘ탣?|l\yQЙ2GE+cv8QӮP/YܐɭO҂og5Ac8.g\i:loEP^|GB$GZ!̱s,`0CV!qv}3q2 .oAIj>) gg[A^ ghO4yjc% *nuIn#M Bp%깃x` \ǕìP ʓآ3sBKKPD j5qf*]&bL qAYjgU/I¾øqV{_֮]/qR^׿kKcyIKq'>%7o;vXB nR ұ %A\O`J4T~3`_M3ֈNV'cV&kmy֨|aT-!~kT^ר%Q^V9g4|C0'Ā=L3Dq^ICd9<q(YJP5FHTSk{,2K{J:,'j6MZ%e+p4x~i(чdSb \%>GXiʅi#jɇE^ e0'IcXˊQ>*O>%RB>5k4:LUwF~kY/VvzSB91QͲA,wA ?tFƈD0fV#5|q !Wڗ l2f9>I~  c pp1h@r5ls$ըiQ#1/Wtvt0 s~OW^Lƈiwp 0TWծL;ycQ@E<"*6;]+3mƭ-n"4w30=J{oG{ڊ2_% xH v #yWq .1⡻!9aǗ/jĕwWq#I.\rM`aOUOtmyۗ[O6[e -}v]Yژwħ]|?5AI4K1j9ud sX;)Ԏ?_?ժCց'c߿6ٷ+beϫuw2\y}J: ].thyNc_cN|jO 5 [(1}?IHȈ˽0s|T;5!ksIfEחKfK>|ײ1]J9M9,νܪHXAsz$63ZK s[MSLjA` T DҞgK>[Y5^5zge83L;]y%uM8~sUٺRZ/?\>"֟|xcH#<+6ZQ,Z9|hb2B֬2mؔ]NMrHT,N>T@Wq|L$z a/5i&"[WDԨ )Fsڞ\P )DѫjcYW+iH*H*$c&ԛߓ2hՁ oBuV,{{ozgn5E4E4E4E)>?k'WoÒLE Խu 9U4Fwѻ]]]ݶT5 =6[1*{{]nP״P״P״P7uNA϶iÔ,_ǷXK֋y]PUu5N렝k2(T1-LkpLA, v՜-ih%U[cv0&wRsb4g_|1ճ9ϵ4gŽ.33-IѮ*䟟x_ޟ3(>C+N8?;BlD@\BAeBy{&[n<>ճjz=KW_ߎ`LH`n.x /_^pfx9qpU,{]}߼}~ \Ly\mgk%Z\l߉e9xbgXugܡ^793.(nڃז4ic'sMV׃oPFDbgdfw<[:s[t"F~'w t_;ˁ:αvF 1Mlri6ZؗRS aBЩ)I H`8[3S)q>!E%F#bS/7D S7܈iykawEG`IUvꇳg¤5`ꢤ_D+\]シpLt-W ؉Y1BLVB8*O"8o6.ک0e'":G݉10wuL7:U{ ]B bJ׺.Vmb .t\Bf+6 YWzj{Sޤ@쁕ʘ>sgu]0]eN Kl<IU$ RJ%Ou;SA.?,ax ݄)iYgrf5պf~y-+v}ʀ7)(ZCД֔M&2 %&澖_s5 "/.|2tJ_jӤE~c#~ ۤM" ep?vn&Q[QF=fDԇp|Mڦ b)o>Y&W<.oUh.b:6 [FR+E ĕ$UjEA?,&SѪP#9]U MP+Zl1^*nrwxբ;p,$#(6e#J'.X"0JAӒIt։͹"D.xBSB6Q|[\s^iQH~T-} aLLO} LO*=􎞛O:(2*`R5 d2 "!93 +r< :QB;di5sr] 0n?[\6-ns}L~? c5٣b9ɬ bǰV(Ƀ7f}A Y'}([)lg{ɠ3,o/&_dA< ̉v|켸qxPbw?p9+N{q'N Bx ZHp_ 4o91<aTK =("#+:0C [?, FS8;^' ֚ZF8XDCS/̭a6#po48Z oԣӭځR Ǭ}@R6].T,xa*֥씫Fs@"`u IܲP7qGAj@=W^'<"iRŝwRZa؅SCuh9)PѨO,|t@A)FZ+Y|4y+^Qlg;uI6.:%3Ng13d9ZU[D6 EAf)%S`8 ΝklB>GEw=QXJu?zu l'B7o~Z ]k=5BZX.Zd\ƈσߨl,fփRd jyLEHdbXQ-R %ae1GSGp>=1-s_3{z Jy˜H[p㹰ߎ8A;oV IrO5qw7cvOfaof;;kU-H:٢<\I?*)s9f ?F$ԳwUȎ V- 'XċDGzhu=ժFX̀̿_iV~V(6VOmuNY\⇳o_\,/&*¬/drlK~ Z/ $V!ojoC Z oSne͎ #z;' *:u D8-;\Ջ˟z]8Oݔ{[jU +]|<Mc>/wrl5ڗyل|sB׻!e]'/E &)Uzp<`pLͩa~"}{ Y*±։OcV\׏J~E͇ji`co ܬ{[Yt:5+άX]xf'*.Z]Cbڑ{ vS'a9S`a1Qtģb|W5QНɦ똺kt;xPW*FS1՘56#Iw( THAE;T [MLACLբlݤQ<\ cf5ʖȻ4Yzz1a9ŧ݇ ZتTabN8J)%Bhs @IEk]-Քo#X=,'?$U .]P!r&@'ʤTTRq1dbu֜U(=U&9vJރerW `濰5rJG}k\EGv%yq`=z,' ħMK #gP9uvL;1Vx60Iʰd&ȎsF(3_lٔ>|MSoOۗ;M7pnӻj Z fsWo_-6Yۅ󭸶ϖoy}Xߺ?zE>OrңbZ{Y$;L֤a7OۉmkRsu<=G4&a{evZ  ?ċPVHc^-hY_nG2YY55RD^Ȯdw;xV['d­Hڡ]ے,EN33iE#DNjDyoEVL9F-+4jf!qк Ni@=V6 DΠ gZc냺Mule%;쀤d !i,Ts;y_ܷF4s3l&@x#iB6ΡwuݾAh2%^Y]wmmrNv/ClmK|T3Ř"cO x[YYAFwOO_LoVPwxoN:<i.$)9jGlg C1tDamU0RmߑsVSq^kZN@בZ^+읤[Br.D8Eu4{PRܢ%Z]OkPW]Qtaxnk\NxR7p[2x lt[Pj/f-KFa+R Ϧ L6!4C}M|oU 4!J9p8*'쑿{yQ:kdI%a{oZ闗ӷӢ.NʆlTlqˆ&QrǼz\Jyp7 8=#f]^UOl( ld?ۇOpa5 :#C>z:8Ina#S#rpa[<'ja-苁o V?}9?&+¾4Vc,ۡj+k7@vo+=,9Dw?|[~(gg~<63JwVQgS"/,{ۻK={ [FyQJuhϳsSKo-j]!U--|q섛1Yphktiȑ:Dh~Rp]ShzOqfҀIlfZKDl8e5UZ?D(_\of6M \6mD,JWRoB8Av7!|s9-7Y9 ITiMfU:kq#B|BPWǍ rwlEx*]-t[1zOa.WI'f˥6$/(f!u<13H1ݧK.pۭGq(լ3Ңtv̛CxHhQxy7[d,k10}w<7E̞q E$\ŏOA?y &nB &n1 .b =Ĕw &s6.SD< 4'(av(KQNG*qh=:(Bx=9k6&A?]flnb%^g*>MVPEm|u>GFtLJAA)>zW G]aMT=+,1gŸqu{WY? F9ix!矜=̊ ?H:GŽسPjs9OV_䃾N:YXzǷ{q`vG$GgFk<"#)BH^#G| d"J2~͇zg4.OF$f5 /1j_ $_4GAyxx2}^?LxxR \V$!Vb г8c<)7Ld~ݎ&ӱơUFdiJc!R<4JĹ#,JK>Am|egk>B|k3_`zB9!l4Y8|C /18gtjosf`?w@K?kbUgnZkqcO:}/0]i޽ _hި\ȶGͧ$mϜ}^:)W*T)0l`v9LX'eYp-v f`jX ̏Ms8Bs/i'x,\8 C.&LX$KIx\zq8rU,Wy1ˆEy-4"91""F4^@{Ċ#_Ǘ݋cebJ(;/<<n%^h՟3O2e`nCjA˛Gjpi7 Ea 6<իhqM@?o~u? O{ۏ8h_z&A4ceE^g\D"vd%\4C_F֕V2k_)WL3J'.돧Fr*'Tyv_/xCx+nl|("ab 2%I1 EiU۰ f]}8ɉe~]U6#3rꃪ}N,ٲf=Z߹DwGOI _GTbC' Ύ9?u¦m3扠 WV !eGDb2Mc}՟Q][9;)8A>R3HjqGa[רǀRgώE-F(՛j`{yꠎ S-G ÈSvr!β=khKD0g J||$Esv wígݝ88>d\X?qV,S\rbɥr]R%϶`6ZTP4MxsMAF<\SdKnuaK#v lhHTYL4 1[5Np3.Q '1|S[3cRoR,J42Br3%ÑbN%ā e&"Pj`4⼔8,9aШ`q FGxuă~1`^Kf$OH9SJscq78o-x#pB1;3ilбqp]R͹_Q8ń3`RM8XZ¥a[yet5+Ϣ.qԾ y XN=HBdTth /ڼ@M =ַAd!kM؄" {IhLr[0d Jycr0"&m9z1X43avU^-f/]]=~|_I~M_?9\I&/я;pŚʯ^t&0op6Aap& -e=&u dL>a 7=BZfp?ZPgl֠!.fn_W^tE/D8B3U)f@מV'TA{1viɥVI+f* %k;> f@RRс:1ަbM=QJVWiJWp'VZ*v ݓiC8+P]mQm Aim5GpAm 4rZ֔cS\̃8,ao. }3XiSs?i-:)}v8L"ё<,x0Ʒ]%,wAgVZϼ@jL*iݻ бtÊ+O-;@xWIFaE B3ٵ J-mn[SjCƺ"sYxK;34f℅~}yH*??^(KIxEzAŏ Qt'1<$ D7!&dݔ3f+?%X#&?ނU`S`/Ia{lmJ/UqiC*9gk_\lNpȲ#JN6F4IÙ!~,q4h [B]Y<粟ژk|cmBO#N%i%D| '/C r F >~[j2ʪB1dN#D<yF\<<[Ͻ[1FM@ٕ΅uŪ絮w\P5kָ}ּ{'CPTSOk~zq^7t5ύ$}JHz"6f Sa9p+`E9PN}cT/ȝN-ݩ6R(t5bUrB:BS(ҰdM0=f ./yH[I+CF*ڹL59PYƐ&zr Ř {JX J ;>@־A&k_}0! +8-hQ1]fL$R\q%}QyâKi."Qeփ{ B-r?UfI,dmz}B7*EnDoH㤀l$e4` n+: 0AUe3E{MG*Dd:gV :hЦ<#fI!A&a*J?0K#{X c?oDG3˚@>g^)YCiB2ߖJ_dYbd#s2)uizOlty1cZn7suTD0D% 3h!%d!rդrU)2J@P )%ɚ G)83SFW<"W@9.ڐm66& K[n,Y+x4vи CmHND[f*6XkF #Щ t7%"ɉ@ \Ȓp`O  =x,6Ul!0 T^[ etOΚ+xTm`Ɍ6 XҼ Y)[Z#w3̶f|`*8(JjqS D*U8""=nYz'l~U[rM%(ʎ1 "ٱAu#J6dKkK g ռa]Ɛܼ-(6Y %H!ƐD3PFKPy:EI(Imfd>b*ydn͛I;O*vRW&E~}KQQ?~&=6HΜFF̬9L 9gzwyǘ&S<8(X/U+Qq4_MHAeF*h)޷>wKdf||{)*)Ĉ&[S&67W BĒ;9s Z@~ BrmrbsLKWK6sl2QwDzsJJkգԪ:65&&$EOK 1E-XH02@p<$u<칠ybvFZ~(PcHik )p?um0TBDo–۱-7.j NZ;HXLFgۄhH:q"o^W>miv{yU:Ex{~|{ȽhS̡@D-F@1[{*?#oS`\H=KnR gl؆rJM3~iٿIysrS?Ɲ#=bCJRA%̉kL#ょ%Myӓy"#cɁ Y6L"P=ұ*y)PZj%۽m ^#-';jH\DE/_J~y1vre \cs C(P0MBֲJZ4NkP{"q{1Z%N;*8e?;0D\=T=cçYn}CF4haĪ k8h[M׳kep2$ߡmd޿pJo(GB""h#}%x4 v Q{(L;y1bD*d*dX@Z핇Q3EGu6eQ0s(|9O]Èf6"f.&XBOH%X2? ut@X.[DO?|JP4/8l>iz vj$B^Wr[pAQEl Fq~ڴ# Pة1by:0埂B8c9–ZJ+ PТwKKz=!3r;w;fEp$Ȝqt gQ\Msa IgeÔB}ҙoݡp%sFF `)#v!edH`VNހ2vD0r;ԼaMS e /[*ȟ̄& :Pgd& r.\2kzBD% SVMJ\?&^'wys0]!%X13 jy9[[وȝP$ǎ[T3ǺTvo̽@6ˤîfN fE.-wܜJ]ڢiXμE[:PMZ?7`@!?j8fgh`W#;y+=T5_pM^ѻ&7p!҄⳧c,~Ԓd+}WΝ_x{֙'=q~Yܐz.Y(X!yquq3RuѶ}zX]!oDfޓDz[__jl2)d/ )R`P{\IŽPrQsNا+-.Z:NH e ]ŕ,[ |x[ d d^xyׯG'akr ?͗ O \59WgNi#vùOh3Xդ6 Rt}?И|g?7p}fVGJOA_ xpS< H8u[ i\Zċ8fS  "6Vָ̒Q@f=9;4ys4!b;\E5kD~lP+k(LrdOqidlj9ZJFϑ~/&Q!4r=Ywg( />US(dtЗM&_*]3AaW7d-hkG >nw:LL 5y`-܋`󜝒Ӭo,l/QiaWo!, JC47! nȀ}(Y>yd Y5V{1N=ck[ʳ~bsJ.?SE5kT)a) ]XW&54 8Ӭ`xx/I=F󻫙ЋeIHr̳hŒgA BܭL [Mv dQq5v=w#{}I\:U&*(:GHKRIo`0reQ]o7QU s`Z30q'6 u߸6j-lM"}A]ٙG`ACRν>OYf:Oc8S oQ:ENvTЃK /.*`Y˕<Timkn$u#>3f֜pZ:+9mH\g&C#y$ ke8=h{漩ק]b a!u9L*!݋))SjJ֬2gdŘAYL+ᅲ2|aG"wux<7G@q@GDC/M G=6prJޣFb"rc .ܢy$@݋( Zޝ3j$O7~ ̐\G7>SBu?mͷyyQʛW?VT"K䝶?cDvqtj,K4$\X^9Dߧ͛Wk7oǷwܷ |^^(&<XC>\ ks64}l 'yGRR„rK|+v;9}:fʤ nUgTnU dh8x2 /]ޓDv6B PFYj)%&emx=9?b5|)*dBO~hD2YZ^=]n5OnSgV.ŷg`ʌ0𢾯,3kK*or=@F*~P0e S^Rc 2Dg)YiJtVuuu,Y/\|C,XZ'03c:,,D!'Z+RS ;tr@tAOOV#WJ] c/:\'넳rpֽNJ)L<朅Ksf BN k"Z%/f޵q#E/=|? C`KFg?\rh3H߯Ҍz$엌FE5*U%|m}; `~dS~DGB HUw#\u܊ _&7_l."W~w.#7 ;^hmxF 774"a6puy~{g1Y?yI\ԘuŦvu%ڬǵ/qirKPtjCqUr޹x8JUXuU%w_z~l#La}6fEw>oon\8]q|0KiXJ&%)\%0H;P$_?0 }\#"§p^%#zgvhd&@XͨgfQʴ`D51t0RyJvlm&ùϱy2[wbYG2˱g ʥ~![R2jJ=Xv2$ ^Z$Hq dUg\yaA;YBv1p,ų)2PξĂ;$64N LX5yĘ9\'8eB<Ƶ/"Ō߱mh(QOǽ]pxy0& RP7*9587SYuu 4G,89Rkj$ mT!HO$F.nڀǡABf (R8 ؃68h o,N6r[96vwܸ/Xz#{0ڂ-=:mC[좏߽|gJBNsrq0u>_{zt\nO?=_,W- lsEQӓ7HTqD5Wff/U=osNI;XDPonYaa'`yy+.j~#$)B{hVI$5iBfl ju'Z|sk%"NԠEfbIH/>q"5VцL-^yG\>R kʃ!pW7pw5*krYHCwU‰;3U AYrjZffVHh0۾Y8ˌz1*B sy}+|-]?k.[ gޭʝBqA^nVxE5e#4cr$^0c>^.q5<޻w&&:g  9A>@ŧ@:I_1J;,/weF]1ǘ58c(]-c۬;525jG7# RݹΠOiRjI!FFKmƥmX5x,<ʏaGd¦*Om ѽ t۫s6~WD+F~f&2Uc2F y)5#1MqHLԅG*f` =tnu0# c!|`Zӓ`Z IIBA^t 2ϭ;.!HU^BBO5[pTǬ!{!d R9u/F&rԭ+^c`)0 6z38i!iVzj^Q-$R֭I^]N5[b>ъ3MAT] t?&&aqHB2㙵k&fe +sf1,crQ>סb%$I(XDC.I"<ɻjSWsCXZ*^n+&.aq\qՔNz륿(|'%:z^JfwXFOx /]I)h`Ci(3eeMVָN"lI BF1pWf1 +sAAo7ZP)m\뛳Cšț".p>zPMχ]>shO ̩gkJ`JU(qi`mfU*܌].JK YB_k~4FTF ,*hE5WwȢҩیE-eEV WNw,bk 0K9"Tg+mȨq-H_Rn_ؖa5"DgFs%'j4PU1%樒2Xq9Khn'{4v@.7^ʻM mH[Aw[:c|/n'6dM~?}DX|]GO2*ZU}J*^/]^GYS˝ Szo, !^HGcToPPǦ\R?ū91 VȑK ?S;kʝ?U\Crb.է×JES^P >G˼Sk%r*:5K^2̱:W{}5g10!T,QËb*Hǧ4X& |8U*ȥYLfAlBy3Aʉ$x fܪ*bUP~⨩5"B18D:E`)"xLP7sŝvK<`' j5E'`F^B`G_Z -DW(do3LbDv]'N֘ 3aEX) ^ 4/L[x=cpm5B΀N>$|2hm ^>&xW<-Mz]+pL)Duh㡇 CxWTVa m(ۋm2$/BS=(7g?lOAB7)?Y~U{ŷ~5r+rf_vD;?dss5Yިn Hg[_9p#S8ј~!g.귻EV\rA. Pw&tTqj*01yq6fNH]M14/@/?x*]jmASR#,9GIe %h,\tTO1 _Ĵ2zܧ7lMUjQ=Ƞj77%Iڔd )%'Ia$@_^-foԌ/+yvy\ݙek%P}&OaᶅS[> b(f>&^F2 x-s)xR]7xIV4JCz7Jsrä.a^4MvYcbZUuYGBpn;11= *yX;Js${#*2,~[Ona?X kn\1V\eev7uκ|Gڙok&i0>}7;@I!J\R4 A%dݤ{MvRfi]bVv2u@ R;1WvJ޻[r㽁S lu࣊nUW3hERtodH~f qɌ@`n!w4X , Ґc{tº|v٤jͩ;-f]zS2ٳ_O<7?ͷ>}z| l](~,e!TR"3vifk59U(&IjjjN5lUJJ൰M.iќj>O5? 7[XaROqLqBk#sBtMd@ *Q ENڍY[|D[zJMmY%;58"i Q3t==2hxF!X:ՂH591z5xx!%RnztR]Sy;ZP_IpH*D,эԽ-.Z%.Kh@sV w--# E\y u9FA NYgښ۸_aeOj32Wlz㊓˦TʴyQHJ~!)$RpCҚeK u7F-oDŸPmQ#ګJ7Yv,B(䉚CěTK b4,!쌭$JWu$^͹,{Lk>:Ú^lyuk#Ѻ!!\D:|G JsZ㲅QQE$ıbHjӇt>PTCTaoU@BT9'WjO6Ld{rs {'!5H9uQV'mM4pW Rx#2Fyn[m|Z(ޘp > AA&`G6x.̞i|g̨} Iv\d-} Skc-C$0,<\&}^lOKk/P*斺0Z[rxlȇEXvp@w Zƪ} ÊN[7T#@ظ>= 1fH_꣫BL]"VP.KEXh^Q뢷O5.5 vz4虅 ekղ5{1ڜ/F2= [yqo0)L"x^Nhը?g).혥>cq/,  h 8SFqrNS#51o?B\6+pu#cWtɽ#_5 Xգn[ #?W%m=ySi84eZ Q(4Z>o*<UnW c0WXM*[8=` Rv5Cux{Mv&$6>c"P;'P (﵏Xcȕ[3O{i3aǣ=ӟbzCJ'ݩ_N 90B;nDGS1H#s,TG\LĒSKKϦE7]Q-\q}e?7W'AjC*7z1LuN|K ,0 LbkOyFE> ˡXtVlx;.y<ۏ̌hE6󟝚}7c٪˿n3Ks\N7L&I@]m|{ ξ,FM 5/~9K r4 U7sLSB~˕vP\^bK`r9\_ ϑepVz3T*X`_~WͳDBYE6$ME^ pxc ?Mw|]<8 ve"q.~|7zuSpGMbM? ̓)x[Fpaǂ 0), j !t!Ri\uyE?h` c6*[{ w]nb aKk09InO6:p/)I)A3Ss>xX*kUm ̀A^:GmwMA^x(S T̎%ҏ)Fdm)Lg߽g;~۟T"8R$K=$jNğhp% B&|L b4?0V ql`P2=|>wkcBN~_ETQ/jzBB0"8#UPQF:JR K7q>K^͂M4}?6 =zr`(,>ʠe:6|oYPAO9U"% @*"YF! f* a҈jI2Ŕk(Ua=6Yd=b!Nd!> ҊY;CS.8&|ᖁwi8\ijIܸ GƳ. I/^07$fF$IOpAX<>U{9PH"9 )1t!؉( ,)"PDu(ɨhVX6v|se/#W lZakgV16IV:S0nPfHd5vJv*bݓ*ATb=ATS[%r'_*TϘJvT0wfF.ԙ_ P=FAƈN`1"ʀJ lb>ڹA@C[%?6 ww0.Htn&R@9iUdf[C"8mbA͙sp\68'~zHAP)陥x]5ȩ 0l15I}N9OI "g,X 0BJ9d ˠpC(hˁuLKv v%fs (+3p10n4n {ӌRRŽ'25NYbNYBqn+ e~H=q)NDV|I2کς9D\+%o6#b6R nlys+NwBlZKVZ"I :3(Qb[* K2k.MvIiczh>[p>S5B`e0$%z=:ё;+NoCu sr>詄'lK1y6x dEE@mfy"A!7 d8Y6&$a'wĵtZ;X;oyAJ%BV7Ce m0!OaV#5R4J Dk5In:@X%o01C$tx2v g@ 'p[5ɕuD[ EQ1.ZT&} ۫:bl5e*;ZRs4~}uv&bXnG]+,wOrP~+XRFD~ z]Y_G~MB^Ʊࢺ; ~Y|wףQ1cwFj~6ߙd> D[LÄ}{0~w1:(Xz?&?*Pl,H qWvof2K'5nhΖ1NaҦ Fmg>W' g#RT`vt|UP=2m #Phzm? /C8;`"%msI4{{;chaBG[L^}l" ^[s*6|i"P{nVlA+4b&=]G ( _bPh#Y!DhD!Ŗ1nnS6YM@H@4gauR ,.F@\2)b@% #B-j`'uOBM BKIyds־OB{1i2$ݜr+M$G=&geLA(ٻ*iPp@We,rY%e"ۍVjEix\F*`. 58/O|Rڞ`'4`e\-T7 {2ZkJF!9i+9,>7" u/&4Jy4Xa:)W[[&Wyl]/fIAQp 6Oz5|hD/ݒ42Zނ?*.C ]oF"qtgu1ܨC -i0HXv{̄!)_('! (HAo_/t4ge$ hٱq#""-ߊd=$H͇ŀK{Wьnvf pɧŪbQAP3nGUr+>=&³c)D[BYILp7Y8GJqt詞 r<͞1N m|a % ae, bo LTռa9K zN%F)z@XRy[(bڅG':x|6Rw ih؞2 ſ]*ydz3 gh6$(;  /@΢YyǶjo"KJS*;Kmyiq F{8Sۙ}%g-US reِ(čv%8[ Ɣ!*i-X+m 22-PkVA`MF)k[0҂#GxI{ 0㹐AxENh$!53A:mNJg^qI:87T;]\P_Ň%!T4/q2jwBnnɨ_KKIZ/ׇe!xۧt_چlزyYwB-C[<?#'P{Y[-(%~H+9_s\s[My_>XhfrJ>+VjOO!ѫ097eKڍu͵-ۭҖ JҖdb:A=BYdDhgdFN0rJE1Ҹ2Җ^)X{)JeS57{ՈDER9H -\d,d8]I wQL2M6zE$ak$8dXSD*T@Ʌ`]D$Iͣ^Vab%d dWsQn~fec[VѥB(HG} `R ]+cS_Wf8dȅ 6:]y 8@s[!QVnI3> k 8)nK=ɟs&z8.TEELə&$v57Dj6:-#ead;Z^]¹ X!__)(i'zAj?sbŴw߿o`?gWo}}Dm &]b߳BP?Rsqyu#o߿)}?'Z潻 Ao>ޠ9/ѷ"2&5}tO:΄!INa7qC 9xڠ6 ;j:y٧G "MGroȕD /=qiztߪL$RBo1}}]rLC~pAC,9r""mn!U fIb-BsApdȻtyiE"]_}oMnή\2cLNmx%xNQ\M"g/K?,⑆?|.}PTnQeSD3i$ THd ǚ'bR<$Z 3ALXNqRHJ}ߦ25cڢL{ITMdeZq?0ne0c"w !9䎩 #$7J I8c?Ϝ9ue{פ5_#`?u6xeq&!׏Gˇeyq[1,R1,nxW?DĵQusO ˓) (6$9E$UG닚=cu7'5T#c~HSٲ [drl=A֝jV$q\hs BED"3ZŒq.0!FDXRZZ?J%tƴ;quV:J#hzזȅ DAR̒0~j`''= {3tT^,ce9-\]X~ȅ¬B{ FkKa¯w n&Uxf s#Ηnֳnխ(gJ6>|!B6b;v*ݪj޲4pc<\S\un/:ynW 1p[5-4"CB<96/*koߪ_Kgr2W9T!n4>lY3 p`fRHyepBHj=_H9)׭g_vw BB4c`xX-uWݘ_1+Dc@2o'x}W덚wGs|}ut-ChOQ7glMpGgtfG1aZEXв[lT^=nՐZ<ք7ZP?l],Q$砦o_vp"ef7K3o)%{T+JI=[wJpe1+n۩)׶4,?.\A+46!pK!oV/rb4 Jl=Q\e4ea.:sV5 < (J4RO !WA >fՂ%+4b(֞BAْ"S4x#ou}li$2qmI5 AZf13@GU!aa%&U"):4!ڐ Kk%F r"dKbI}--Mſ \aUT Y^h >VhUJ8fm&XJMpG[+X=epm.ikܵx6%­E(U M롤!)%9ryާKtht%tAfEoMq3 yA LH:;)٪WdwŞrEWRB w Z[ aorCeT5h9]|݉{>+,`-њKZk.![rj9w@z( U^OA q Ӌ/k~קl.K.HtUɔּ~^ǒ{m;6#嗟嫽8g֦ YjwC-x& sUɐ< ƇB0B"iѢ2$sF,?6B:39'{Ic$2AڐM2XciY6 i ձyfZ4$,hXȉqFPUޛrf`4%2"O13Z*˩T&Li`΄mc B4Wٺ,55l̝I][쁚 'GQMiEqFO^$8-,0;_Jdǧ1[   dy19ncU&S1I ҊeE1Sa9yN4+@In88H}s5W3aG[9F?*FP5]7USyI*hRs+4zPAçL6qehԒlJ(UZlN7_7FGc-fpԯ=K*b;?+ѥj~JCYAZtNϳ:|5ݹ"Ek`-J"'#'9!E +R"&a^4JOW56c<Ult:Y=T Vx9IܗwZ47:lgJw_3v^|{|xQwIЛGsf#j,<5Cdx蚟wk >D ]y@[~]e5?]׼?O|wWw2]m+zȫ,ɞw9Agr*wd;'A&l_]= rL`26_sw;YWUt"j<{O""aU!F'Aoވa{S"Ta]؍fJѺݍ6zI[38D]ݍ((qVzvgw7B'+%n=O Q=ҍL;<46MO:$wap?rYؽ+ `Fi[e081}A˘d+eZ` /c+Cd4RNPsAe*gHLzMnVHep#`@c=LL'(7TW`Xf 3: sJ*ɬL([T3ipFm՚Z#SB"SB2qJHTm~HWN yJHN h;oJp4nJpZQ 8O?yEHO+gvoûjJtOWy߽o2~WeeW噴 gZ{7_e0\$bw0feFVƖIN[mQeU$E4M:n:hY:#:sp TB` xn)* $0=z)HDž9N3N%L$cSJbåg9ϋ44#dZI}9Qg #B#ʹ5B2˄pF:*8/CE>x[eఫ|xš2m0cLUΊ ˵|OJHؘ3{z=\E{oX>F!M+~ Dk?$~ucpeQD6^a_6idBsHz' &F9;:+fu֒"jFホǙat-$F0UP1 \˘5TE*`CG[T[,ែ 6&3A;j@a^3 ŭO Sj)Cc`1BuCU0H*L\SDBT\Iic-884I|q4FJ0DZu ;C*9&X1Vκ?|'v1{121c Бs oR HN #X9"A{/-r+vd /H>HC=Xdϻ֒~jWب,ќ'&%`Ct <(H*+D81R=1VE]Ę" ℓt.7VTl󹀝"d+w%^ ;߱#6U;{}0S身0zt7y?fRjp懫3xM-s+vbA>!0Ǐ~92溳7"Mxv*aQ%ӈ:k*/1 UD{# O#LM5JQMaGlJ:k**6@vap)nxF*_l6)/S8&[',3`rTFy@6&_]o~YNjL8h\ K1Lxj?|R(#,!K5Ͱe0SeZ?%f0c$pfwc!Фyw4nBՁo|%1|馋%_3)B0'b6M$ǸBjML'4~Pff`njV._>/:`MU[OEsgpv^Yv$I;}Q%?uj`7LKO j RQ15KORF@S٢ͩlLe!5XURjED϶yEQs'Jʣubvigբt.nkUDBR١J"͌ +hBF˪ c*_B.5x%&YL9:.c<-!9=^4Nz~3_eyq)I3em=mY p65m9VtFIJ!Ǩg*|5J{OozM'LJp)c9w>觡í;iKAU^ T~b/&?D0oC6D0oC|~O3S`/5~+l7@9ErK "9Vp,evShU& f|Rr[7FLň!<"נGihyK2Yލ&.'wm1>TxS$7X<稩;zK.h3~)>T2˽4AHk*F.(3y*<)W:z Z6EcZIÚHnװ-VY[J#?EȀu Bt4hR%+^*r+5L)7m:Tu68-;e+ lR0ɽՂqs3k (1D eJZX38)nCꇒ܆ԏٍ(XU %ȑVNV cd׾lc~:8l =v( ۇz?aнoiMhkHzs{] X{G1F!BeBx?̿{2"q횰jZT9^ǣNw8dxӧtL*W6ZsO&0dmF'dV=9UFE}&YĈW$V<96tM!E~Qx|usI^?^<\uPѴʇz^sqpȶ 5LbHV9#[h뇫Jiizh ă_E'A 57be(1;Q t[Bl.C TJ8 23 =G+~0J-dQތwo7EAv@J2$A;KO'è/W4L%(('73aIe3Ŭbλ!HNh&=ϲFb? ")bF aŵat.!>gC egnUׄ¬ sK0䃼[{ Sqz4b3 ,vۥhav>/{q5XXȐafxna[c-82˧ձ=5mT?fǵLr9`( *B19_yջ#taDQKy\ wLa3{,$"Nem&ð6'/?~ >d`I!QSSo91ukP$1_(^޼;(*8\ӪY] 31x0,P,4=Dy<AH;~)>eWW70TOdޑ_p?6%fyqbYQgAWgrz_"TIHĊ0}fѸw+NdO$ngX|8ۚL~p6]?{7MrO)]._6'Obf^5.;OY4ԂL~=-?\/ֻrvs}#(:mnFnbGK;m?J蚂{*"dvGۇIIqUsNEȀp@/[ԔS@3gLs)ofB&Nʊo~6xf jwq#`LR5!pD60 B+VXu!S?]iNY^`f%ۙtУ<ƳJU"KQTzkPUZA=$6LHk*ƱkDҧ=*udcIUe PUBa-OeU%BBJRQG4D:r7D;m50$̨W.ZU)CIJ81fXTQn#suMUw#UKV#HYġb.Td {|Gӥ.CEB|XHK:,ݝ [(|אsר+$g[20I!#;S;y‰S[?9f4VQNP!U;όf4hFWfQ%5 shբbU(L3F6;k-f%jv,ũ35{]BrKpFff}Vf5N׃ewqJ\*N4 HԨ\=՗@c*4LU.m+5zֶRWLngj94} iY7>vТ 2{ H_>U uTFyeK7n0~ I-t15hL7:6E[%!Fp6Rab=t%Xͥ7i:nuΣ8 FLg,'&35ЌrWx_w{RV+o\0RvDz56kA[_ǂzKM~qEˁo !Au^Rr<7I^xb_H^=PFPY@1iBӧ_R@'Kpm֜cϪ&aGEcNe qK"1ӷv* ǘ5hp;NwwQ"Q̼%5,*Maײ%/7`u$=zGA@ ;R1ڌZ jk\0p!Ԓ1 MJ@X, W1*o% y~ mZ@@WL kEQPմTe29;ci/q7^]J >#lYjTMUj'c:qBzSngz]El^-VhwvZWwXD b ܳ_MpeY~9jux{_W^ج8swE`D$'m\zK;)!ـ6-|j7(|0M1!T򽺎G禃IR%iJȹ+`fr$deCyOt&H A<LMiye5+V#BmJO7a"Z_|1w9wI346D-+-Zy@ePMG{c}*t_> %I9 y>jwHQ@?`uђ>e#FFQ^`f!Jx6cK8 OM12lfwq||pT0ηuEDq6{1ޮl1"HQ^cWm[ ]/9O25W|ԏ Dj†T m<./_aCgʚF_Ae1 Czذ4mezuAƚ"9$E} p8|`Uˬ* &nt̰&R^<6iE^CTp-* c}d\FkAm{CښApw0A&ڠ$F ӛOL'UD3p}8?{-'J֖|7_I#oz=.Wm55]Gw 5}tQ!PIJ dݔD/2 I@}y.i*ʐZB;,}]xQ[8\ :i* RS) J9o>:Ҥ.m!-!-I#/舓%W8f5f>.~~kcǶjR¿\•` uqn_'YyGP ZE~E%`1$$`b GI!̑jǢ:B/wDZs\e<~o B)zit9; TljaӂqbA3a9L`"2 I^݇]3疄4Q0k(g_qs/mT߮ɪ.,*xL⣺@LHoZUrO~>頚(nI($5̉8 ƔxJ@3]FTBf6Uӆ*-՞jVn|{C!90g]q)d"_xP\Hk*f$'9S Vd,wTwkB.V&3`$7u6{cDC39{RnNUmpcoa4 o+Fzy&, '?j af&C.ߋ\qFWWO]y bvDuWA'#ub#,fFP01S&cNJ79L'bxWB iExL{n3+PG{L,Qel{Vw7[nw'#93('?W~ӑŽճ/GS4+ruyy'x6ěQ}) 0ADĈƏ/#}%h<+p|wzf<]| ߇3Wі㷣cřBaD}y_tGgb%BT[߱֒,JQ{Ck6n䇽$mD]57"{#k#sŹI  Kއڠ~2ux  %|GU A.9MsI.k8*opW\u}n_*Mn P?, ;.g>h҈hr郺Iٙ_`>NL/ce?I׏Gg~~FSL\. -FYн 7Xv& o>|x^!)+h-XI9=޻*(^#]6Mv2eiiCD6.SXP\_&/~鑷^И'dC%<>A*adane n NSIQws+z<+o<[5s}7 o( )?GZMnܦ1mD_H>y{[KΙ! Khf5ZT9V2TU&/ĤMq#G[)cu8V(l̐1KmBiVc't4}'C3c 3d/ll/PYs VM}RŅоIN|n& i5s@]_26zK($۷s%I]٦e[;5ot]Q m~bqBZo,I(A"VF_Vā{ ŒMظ:ݕ_]OT~'lLDSv#5&?".!R=\4U[)oՔr:.c|̊Qmc$\1.{KJ o/\omy sivKKbCp/A]꺮ziLSI[ݰ_0z|r?Ƹ1%@qB'|Q[43N/ . 2ԩl:ř|N|Q3N|&fa:ØJhö'/kфTopTm|ښ^O/MSr%ry wyBZλy#Gb:fXp"Pc> sa6yA|r]qQi`dB$ cp Ifx¼ 9`ښzdڅih.%:Yel~ёq$H̺ Jb=O*10t ,xp@Z3nF7Xq$ & `=p"¦O}FFZrT {@ H#-3PaR#@xꗫt9OSLsD% j2.;@$%e?NFPk(+:G? Z"goL\u&Bä&rpry(3:#lK]xolxز&?* [*LڠA$1F%9 ,5(*EU"SVD<Y$%G syt5}W`y\Ë`3#UWL|]bNJT^|HID$b&XPn8L1AZmF ITa-;Q)-;yqx,߅mLqTMxc1^" WYop뿦nM+6+ ԮMZi>FQGL{Glj&ҬMvܨ@t ,Nf&p0&$cN|DVZɟ/: _Ȟ*Uǽ!?y;gOd17`29FwF>=zn4"+>𦅺뽾l# ȇW:nmfnm J&I81Rʩ h˄# I JDNK,ř8;)Ι{&u:Vlˡ)CS ׺>659u{xާc|+)[(۫X&8Oۤ6nImpmH/d E{y\'ㆃ,$ 9l{?hALrr^EЅ0^3Gc!T(\as]h1Fpl =!6X8 fs$$8)KE`L&% %+K-w  T;WZC12$8w%gG y)[uEM %}zՂiӹ{Y_cgчY3+PG{L,BS7+ V£X:! ]mo7+}m,Їlu_.9vȖ_qf$MXcN"LOŪbSHy MP $J{eoZV:Zբh-1jt۴- a45Җ "uvW$ Ԯ ȶU>HIq~JmiSn,>ncҚ,/qseh8NWK&թt$iSØJoK1 ٚ"M"w垫Ex[hщ J %tҌ,.]oorH-$VޔTƒJJTTZNFUCi*W)1!3"U#Jb5)A8ܺ -ř5XT@K5Rp!D %mrRt&kzk]8d>XEr^~=VAH =[OU΁[S1BFv:ʂA]ֵBTC3ZU RM2[OZ~[]#a_&8՟ 1[_x"%Vɸzx拯xu7~ x(ٻ^]̚jSyQpu!.i~v?PP՟()X&W8R%P 5!21E~Wm<6Zx S>,OOQm Bn%ծPՅO.|zuӫvzuPYT?]:e1eYV1dSQ&H 'pX'PJR (ڌ^j:z"=qIG[|rX`))S!kzCwLC }zY{],3H}..LB P¨$Q>yCQj"-w|sSWʒY?^7W! 0ZbXւ ; ⿓St&D+FiV3,ڠmKkggPU_IkF0K9sN֮@MHhCK,6(;H[%G4LVeZ[@WHY 7̈ھՍ$Ug? _YϢ]\jd r],9!&>Ѳ.: &%H<|KO /Bzbx7LliPUJլw1CuX*k!jk5tWSa]dY^t?Io6G퓂F}saa//n4_&~f[|͢—^]UQO{5_1n? :Ems2?_|Y߯V] ˑ }뿼uŏ?L(C?}+F!Q /P.Cr4OCK˳B^MZSȩMfd-69k$ݒ'rD;)q01o6& ' /bɕf-p5(CReE1EXѓxBu}#}W Gx=\{1i I4qlz IJ xu6pdC0U gX=*3P20rKG[IC)Ancr (N*u:{&ktJsj>Df\Ed]m߬/>C!hR+ת;Qs0ʩiUU; D]QRI( JqTN.Iv8d+TnΣ"C8%i8{z1&,zqi v!(ᄕ BHʨP;A1ZyFǸOp0>ZyFǫ[`8sm}rDҐvς 2 c`e$J)c/zS:̸)Qހ[iVuJ7k3yN8f07f/  N; h~`Ђ1 A6Rh 8ZJHp9"qO Ot$8֊W T J wgUІ5">!a'M5AoVJsR3FLk$T%8,ܚ2 =wkߖ?.f_]4f>Xe]]a}Z[P}x *Q)F_I7 BI$dF5$QKpRN osţq̠9mmK|p ,%^k83 u]k*m͖mFGc3nᄍ}I֌9'wq65QbUi>bdFd6#F81T:JSj#9J#+GО"*6mnkpȆVX)ۨg=ŅzH)$k2<~L%Fvf끯KY6MB8;Sph#Ryw;VnOg>Z= Z{J/ PD^%hhG'w}T}ٰ Cِo57 iqPPas\rP˯|H++?(@̂@h|=t{ A>}1Q~Њm朸}픱Ռ VvEl cҐe)ʬx[;ehO.1U \_Nww_BI"B[ ^;U pS[r,k,I-,宋%J7.v"4v0f0-~J_ җޗxܥ).ňϓ+'ǐ!V |?_|HHˉSzVfpe{*#$l5+ 7WK#pxY`:`r-锔awgc k[7V:9Tdvq٫s 8-yx˱}M@< X^tz]FJf(6> +&:ܣ La6SHl0[?I u{X 1IF~mogOϬo޼_}fV;[^\? ߮|J///S& gG7g.{6A H<#*2$=j/UZAz(ʁܠɝ.{Yv'ֵ-7UWvc72j03!oJW` [["l˺$G@\Ԧjkurmg}3 wBR93V:co(oo>U;6c;6+w/\ڒc Az%Uk _iIm~QV.{]zv)`N-*QAʓigb>/ If'BGa/!vf{8{zAA'>h?P&i!}7ܚ'kՉ*h<$|qGqY:kճW?{ϯ?|'ְ~A~jrit&Ff{LE $QBË"v;қFmY%1e6UX4-$Pc En~{{[Lxb#LSxoN$x䩌Aiu~w!tlmݏE2`ܸj$8ϞazvXv*-}V_<ދo.}̰T|6_j1:i=5҃5Nlc%ZSmi7nvK d餾#F\i@c%9thϱv?'8'|&τ;Udr/w:: TVUp'TcEC-eM*v0d"S @+IRY4AIFv2 dszN^@pb Vd,Dj i`;e2H6? /Ϟw|b : yNVBl6;@q,Ek/G =uO8}ɌsӓV5 8Eul'ȅqgcCM7zBdKr2C"xȇW܋~ N)Ӣo6s4\\@Y@YK4:LsBM.$P!9eTbBܹ)b:0' Fy& Xj)fP#CCbƉ H=E?˒yTOi>9 3E,i,QLMDRM]­o6I\t]F6oO"ԓa$mpI'!0Uyw8|=LTqeLNKmvbQRQeQwmJг Jxȇ6[,-[,E QTSJN4Iv[BY+@cɴ4|^3Lv1d{TҝXƘvTAU3NcʱGȖ#(54Y{ W04 D;E9$ t6m D#A#,)fw `"k9S 9)b*"i 9)͈%J>+Z8 ީV:xK/AJm/1IȄ%$CXsp>U}\)8_ I: k2H=\W<E66LS i I [}v].˳;ᮆr"eB>\Xx8KU>XxN$q֦  QNT8#9J$t9cop2Z>9]gԭ5"Q L7fT3Y'<枲” Љ8Obgguo_)˃z[d)XϽ %i 'qN "h7] N"!Bv3Ͷ]>N%7s)XC|!}_n:82gIu G،mD"ͪvH ӻ?8#bb} G0NLHoU0o(' @/t)JEt:2lXoOn~yywX/fнmTA:?cб#* wm6e;ӴV(^ NfV89^&f]NƗFD{!ȏHb04푌q HȏE mB?" y%?Q_-A/S>2EÎGQDGN6k$Y|ӶΈLhB| a$S'1mab`t0>0 'C"3s7Iy#G[d?$eL8W}Q>w{29t6ck78rb@PDD,e*DM*W}N^0M҉C:^xls\tF90jqJBʌ^H|o}D0[yLvuAv+Wf0HTH'lW$]Ĕ=tnmvJԽo6wi HG 2O$_n tz; 9m#kvٮ-?|fwqVހ0ka ;2ǸasiP%h T;WfTO2ּ߿Uݺwtrjڞf{=5ymTSѾ$g2]nrX.S3%OJ|J yKwAMg;oًw"Ɏ>K$b&GAc0JI({T[VϙײQ8 F%p#$8ؔ4g=QMˤՀXHFdMIȧ"YF%$]j|yV8G OobZqb,x2Ɠ훪D!=ɮ.ϗuS_㽞4χz8iWUQa]nۤ*ڬ*Q {vZ!JQx>Jnl`JְqGFVDXu ᆚ qΫ 65p.K%MXE6)o{Tsahr?#U'*:R%ߪF}qTuK"*Ƀ1k!S~/^ (ef gV+Jb/R( @W𵁫 v{1yH`ݥ̼1w;h Ht5+Dw9㘵&bxTDQ@(!FAQ_#*VRO1 DBr_@ W6:ٹ%*@uwܗ R<tg@ˈ mg( ;6j{~X r˝Va6Fȧp#ԙOfh1|=jf+jDxOghaǻH$&NnڳWLNjO2WK'tc3 nb#pD '+hgty{%r3]TiRA0oI었8I:JB9eIE?Kv2/hɁ0 9͝D` iʹ@3[ Ј@4V d8&pJ*%E|$ x!"Ԕ0Ch` ħ( hشzL'%*ʀ1R9(&$^j~] zu8a2zK;a;6]&%;@dђt:{R Bw홁o}ir}(֗ "}3[*+${mLYƺiyWz2.?>6ˤ#`&/J %OO(2hxsRdH U( 6!{3I>$#@BCxJd 5K\PG.$.9$ Ҷj_D @J3ߗ>$鹬ԧKR J`t6 \X=3fNdTSPNz/iS{:Uڀq3=`X' 3qZ3+w?K( GSh[M,̸fZgjF OarBp<Q=5]_N}u$AX0^14Kqm2ZQ!E?F;mmgSPgp Nìkх5uO|>|<;L"`A`ecÏ!eߎy+s~r;UW+[?MGIxI~6,A,Xw: 7_d7wOo<_׶no覄}MC믻=J^n8qhovd5͇=;wtg///?soy_yƏ9Vi2N=޾/wgMd_Xg="3d6SڡMf8OK}'l睎h zZ2̤s{mlZ;mxj+rWǓ<~哞սގ>2^T$"kʆA<xu<4×~7/?ڼM+Ӎucp:^~4yA`GwѤH϶Wޘi9[NS3A=U2g?Q]}:-bkZؔ}?x<%{?O #:tN?';9~N?1ak7my?/7@ x?2]cZ.ʽUUɻ:O0S#E1Cڬ^6P0 gh߫Cc5M'w_U9x!Q;$7 c`7JYu-S[a/jwFs=AFϜO*,E"Jj?6ZX_;pa, ;Pyk:%ײDC33 (+:KB.N}N/"euDw! & Afp|n^i8>®Afs1j 2[Q h+ybsP2X  KlC`6@AQfra0=&deRc6t|@|l~0> CyY@"Q(h}!t8( "2V LǸKhʼlm}aBvk}^k{Eۏ(2}kSRl5̑v*!NK3}Sp`"-@PPٛ ֔g:xWzGYH{9skj'j)*5{ЉyvcB pw2ea4azY$Y|9?; }qp*wmoν%\`m=iݚzAtFA=FMCS"H (bHACuUJΒivkȼ[pY_6kR)DTŘIFq걋o|D#`´ZgggK*KJ@;(aw\0 sl-:q^߹0@s>ocndx/ĬŀbLЧ!AyB0n}@}_1 1};k7ӕ 5EpK^kdykBadO%bS%D9S3Axf~־`־I* ӝ+O ݅.U=Nlx&J0i->[ѐPO|IUF_#^&mDʡl:ڌّƉ}P%5ER"qDC_UwU5BJ/F泪^y7kIz=1`.RۘeQDzoe7~*%Oo"@x)"m| +б X}z>Xs,$%W3_CK=P\ 7'ŘȴHFN^ibKʯ5ghy?ǚ KՑܛ>[uS1NoawNO6^GNq{Tj& )c8BqLP,X*@g͝:n*"E)+ r$uAh^Sji0Թ T0lg+Nn*‘eq&F5 !m绸u_guV^guVQ~ִXhEȊ`@84hox*FPi%LZL{30&E髳& K];0Vomc~wVI;7YJ$[UB`&Xil N llV)RZǴ*.)N6 JFi}`iLi{Vzsz9$wV*Ab{w*wUGUCۊCɳ38O0VT[ BYz&6d q|)i*i):_4rq9YqCNYׄ*SBjM 5!a 2*M(YÏWcPb,2hQmYLg8S1pirF2ZpiMa#BT`T̹Xwf˙E3) dHGs쬵+ 4mI|p & -'nyL 'ܞWt|0KbNFc A睶yc={%y+w )!"ĝ2;y9Y|= jA5p ١HAl=Z!A&)$ JV}ɊQuæGp3dHuZpe}7)f_j|vPiٰ_ %\P#;`:b| p&`ĠACR7*2g6F-UBSF#W8=A̛l;\cle\gb{ĬdY8?!dÅ,Ð\iݟzNt:;@ UQBgl h lU1^ ol /5p9߫+ 0} ~haϕP}ahxi=J󑼜KH/(1^Wu­/k&ɣWSe\RZRbPO`%F%֫<(ۆI?$UIJxKhrh2MU^jq =-bZ&R"@gU8l3l+6|u5Z.ϯzY\)Sgs2\a YhBqTCULH0O<*(l$m.JuH|YRIڕi+3w֮ԃ@sA< h3+xZOclصΏT;x'Lt*Sr\FDvꗮQWOzdA$Fd숍_Ik7{!/kXwer/F\wPwUk.G$9FzK ]yFM ^b%UJ2+*@[ Cc6t-Gmth牒UTUOy±⦄0EʣJ…!",y̦[`n}4; v삋- ڠ)"%99l =9cdbt,ƒeuzq^F\kX`,BHN"x"tۀFk/} +bf0>ლk|n]?dvp4eOr^IehC-GS,SƄ;>kB@8O&DBLQ=D@4k% 4-V/Gi_ϫiϮ>M>kM&c#mܕV#(/NCE2'2&Onmm 8[0vxm o7P^8AJ_/Wuz܌߅d>_=|q4~&W)IGUtd딬tSu9Yi 6i4=i8Ma gĪrQ(tNQ4ByYA9dz83ŚPHd"e#oKq56VݵF5õƬ(푲tJBRj__t\6 _z %.:Cv bd\gn9UWX&ɤ^K<},r|qŔsΎg Jqrf뵲.SQX;`S %z B"͆о8.:#v| k3T AQW/MW2wڟ#\k,O(Ԑ Fᗐ$QG_ݨz;IMY6hD;|{q0LM/!>7_%3ݺ.g禮D wdrQCy6:6}#1 Ԛ%T"xqG~ƒatLܽWgntWg,Vȍ<ԈhĞPmP~HQ.ʈVPT̚~:"IIZ޴./yFRTkjr%P~Q|4Lt)v*9h0/WϬLq)UĜJ˱8 x6 {=qH@uJviJ{|"u%3=HKB{6/}^V- ; >j0H7}-'3roO IFUJc? g0IL1zq5^lF&IW:!XMX߇0]Aw7םL} %Nvr7/ÇMBq-%N!WMVڗu[ib/*ẹJyA)$ēJ=Mdխ$Uo|}u-,j::VI FU@Ѯ[[c].lE"mMBtàJa:;|*$g͟b4u(X0٧yHEw@?ׂaBJoo$-eO?ڟZ" NkWPɚ\% ~0k10#>6D[? 7}gqf0BfG\eiY!k~V+v2?&,'iV2(wDHIk,alEVj޷_bt)&an$p*PcPWQb ^R6@3jr1X XJ X"\Sx7? [N$8 eXYN!2Փϒ x]XqvSDwRuӶmwO}zO46D@O AWBD|K&HkE֒ur8/T:Q{o8Ii$E\_gf[f!t<bq3 HŧHcoΉt#{XyeJy?؆TRϞQpE_nmZf׾: 8BMgxQK1#}jzRϹK- !Gߕ\^rkL9f9Z;z|vba/5Gs Y $B#\ 8yF&#H#x$KűL- :O*R4t/^}޾pz; %8~~>>:j7(Iexmj+)_;9 )(^y0μ:H5E93y{a1 kH@=a*%P9Ny)tUDgtޞTv|ZN0I |_(\s. ٚ·bDXXʴ K{ʜcM, YɻkmIM( -ms}#8s/)5WNMJܰfѻZA7O0@[$Px>ӽ LK"(]aW#I[ɖ96 D'{BO"ȸ5;uk2]\ʨHLj)8P,ZvA=ξX HݡGOyC̘VHȩz]lתR'sQVu* N$Ly?y߂?j T{Bˮ#TWgGC؆08b5BaL2qf^qfؼhZܔ.G.w\tEoxȜ쬉|]:UQ2#ʩA0!Gq) Q.9RR!#/{&(.VUPS(60?鰵9^ٜsAj9 Fat4~kf|=Ͷ_+JwR]pWx 3 SZTUבQlDDrΗ bܮibCn@0 !r(*Q9գe*א]![% kd<`[ iGg+[>=^5a0 fO*liҘ:Y{funjCrg`L(Wၵ2UEkko;[սfd1&7Y̼ʙ7)ROuMڳrAUďd{D&[AMқk^ܯ>#w1j]rP.a\8ݴu#;óʧ+M&WPϸH2ћkgn?3Xa C c;;ڳc?_?kMrgg><.ʥ{4fhp+gq,%jWTe<]5oF :[W#۬[86-(%=[o]w-~5cRbj٪YjuPO/(IO?Jv{xv6gߊ7*&41]# &AH Dl%ִZRo;ů#54?'2`0Vf|]@єh+Ȅi%ppyʍIJމy8\֖!rH)dsXLpʫn'  I%l5b ݨ:`&jP)7U\Lg۫ %0'%䧜9isbE/-MC˰?)CYQޮH Qh04"yy7U>|0VO4' ~oY!T-?D>ϧ >O~>Agf>j'b< _n§8+ P+g>; 5<磀dcD` ߿=xĮ)` N KZN#Zc6͎EcӰQ$X0&6aM aYqKe#qar8u rws_tZj##\ˏq/ᤐ:﷉S8onhhnMĮ'7WA‰ׄ(@ˁ`~rC8fNX.1.Z<̂0,%`z >6W= ݴBB,,qy,=PaX!9r}1bd*Qm%=nzih2L|c Cd >qgz';i  m0Y?eK2֐7w[ٌJf7G/cC am.1eNsl1yT,%4?a9n&"buZA|i6fq6:{OʎBvhelz;@v<>B-Hp{>sT5NٚeJEKp\lP0{q u m]$XWŔRa8KI]x?ۀ#Iq/&͕,e: tMϋTMφV i"ɱx+ p/OumgxHȋNjw{*T+-X>fWQI9 h0O5X5lkW@Azut WV,C[V3>2U{|WHt4aؓ 0(d1,uٻVn$W,vgڇ,x9NNaH`_f z6e%\f1}-jYF ůŪbUѻOH¦R b•΄#0 $ .xR.'@MF1].@i|0f1Ք~\}4z*TF<9l պW~ ˕rǰ\1_Q9K^X2Er&%2A1˙>\˓6m|5zu˟/1#H>VI:7*ʡ 3 #xs0fJ ,klI1CKGu/ ְ[vkXoUDjSTȂSXƹ . \%0K< E:Gח?[>E /cXa2u/ceF蓤# fLHhASR&%%lt|Q641ڹK7hSG2חpO֪HD`T13>&W$9R%dLdS dK(L'W"q\[F)pO4D<qUӮs,Wr2RM !Ƞz? zȪ5)38%+;]#n]IF%(>EoE%`$/C 15.k])0dvssuP>AۢG}v"} UfnL_^iļ &S*t4LF71%GymbLmj&G)yУ xf:wCM=ֵ}#8_,[qĐ'{픲d̉W6DJyS3!q-d;'͟$_X4bj!I@q}eqcWYiF t/kě]i˵-^67qh;NIZcnpl7 7,պ]ڇCʭOլFok3.Z\8*aGҴk!/1qڣg: xĞYhV=|pf=AabxȻg^`Y-_~hnii呵{uI{M^E[N+Fpqt}N_a0Q-ZӐ#OK5fzl3- d{̼ vwDAwG\^gop}YCNI8|~OHf:c8c@ jMl X:۾FE9wL}UΞq ~uoLXsԷpSN&:[(gtdQuiy0pD!+^`n2KkG~ƂX t)sDLυHP%)tpZ!sȢ-+ċ }6XK[B˴RLIKJ0+s s;v0O ]@p4O.'uvߦLǍٓsOˋO 3 1('bI&! Z]ZioY {jlNDyrDKg[5ɞ+mᙤ8+@/#u4jJ:)/4s:5D)\7ΣupIU @ˢf³d9廰k{[P#ԈIf芁pc8vtgz~&_,Ӄx{54项F@񇇯PՌIswB|e]\U1Q-&`'F g nԙ9Y%&7]RWpѷ7`o6Ji.=$@nbtni֭gA5F=_zEҳۮ,<*A{,Bjb&>G%%*@.Jl־GT-x0%}J2u?> nnVty%4˟Yz?OV/oiyM>롑.kury@?lѧ/OU|5Zb8(P%dJܖm?.<$]W"Ւڳwu](z4MfU|.$բ>}QB%G[j!rV_T5}~=+MƮcm0VTiH}Z+|P!ɋ;1vizsxX׏Sr!c6,G@moFXK;YNHv@NM8n0V3;N\W_̘-8qF 8?X,nbS8, d-!)ȏ:BV\!aB[6q}RsDHRXs_}ETGSP#(^1DKo%ic+`hј}RW K&K)իyƞoAK{}U2kzlܰ\+{鵰N<5˵'yn[f/ݐH}.i]=Q r5I^jZĜQ'EɃNP]#:N,(4*h#ؑBt!\*J0Qh0 `Db o23s"CҩW Z}ŭQa\-Q]r*.q!G+'r3M!%]KMBŎYukU( d2 sTlA9ms]гTٸt|sd%@#uGL#8W0RvGAΞ-;+[˴ ѥ&+d"uj1_yDmH=gؑ@i 7.~ga>XlI\nnggggq7>'7}A< 7'H$CrYv)wZ&G:~ҏe/幍oNO5.ԪŜU X ` |&qV-q{bMS2/9svtd:/^^aJP?<Ң=I``i\.4]T3Z R?߇Du `2"6z'iFռa{v63 +"qi gڋ/ȽE3f5ux\ܰ "gxF-ȽkUl̓<˙)}.kF+BpM/cTR/Nq=t=}hFx4X}47}{05'VWR%[92lL%!cQJ"C=KUd0̹Q bӐ0+2G [ L˹O NG@Zht2娐g (Q4Y:f/Q Tѻ U9aǝ6V$# 'm' %6AIYBsJx魳B w51hd3!Rh;"(#sN)d14$)3zn%/H L#)XN;0$~ˍ6ߟA|X~}KF1]N4}uzjɧv1 `ט`nisaDeE$"y1J'^efM9N% A my' ˩}v٣lQ*sd aץ#Ƭmx-Y9q$٪9ak3)Ob(͠RH۔H04yӘ{mӵ @6temw4.>a -"z[;_JxT [HMJ)J/zAZ`B"yHSs]1r7 6F3 3Q5Ĩ Fo2,Mjf!aok\C_h2_/hǚue2F. R$EgAbLx,:sų1!pNԂ5oN6"f(YaЊr$ً U=+]6SY{IMKh~$YJMlH\R^7t4 <]et]ZvذܰE6 W_rdq0,u u6dC% mB7dei:H&*/qM"S۞NRTV|*qZ֤ٻqdWbV4 ,9iLΠART,'3GS(iI,SůXŪb2uF`V+OE8eޚCLBKB:hy}E[1j xJhiaFYu[}Z.V.[? 4feH}} aw:,ǹ>z9ɿatak t]-̌+g$H _t *#xɝCpTzgaYBޟX O^Am<.xEsڑJ1|Ղ@`!1o ⺘H!En M8>Pzg/]n C&&;zvw_ޔrD? ,Vϒ tzqx6pM/Ek|82@8 xv8^|8GF̣ IH\Q(GC:`~*-֊y;z y%!mȏ.}Dt e_94 kR$ Qtc ~3[\D? }z7(0oj>P&vEDF\CQim{s[n `Zs(!Ss [b'+?93KTZ(ŸWr"p_>e//Zᒵy׫4u*7 ~ˣ :hnT$1gR>)8X4xRLHsLN3Cek8삜8A*V6=Φqd{<є`eE%Q|>q0&,dE9 TD`,m.N6ViP0ȔU&pYeiqĴ s1,,s{&gdcBc(DY5B@Z`9QTBRg3T"$"9idQhBhAirj gi^VwPe00n,Hۧ xGchyLWuӚC<]jTpM}wǀ>,p732E2JIk`[xʒ{3AJܐeTO6 ऽ^m:ʁs !nulJw-;y#l z~W(ҡC40Z,۔aY?$$9"T\SG UbdyȪ%',W+a,<ŭ3y-i8t\L*_zzwc?9IVL-GY( QaY,-F8.y3hJ1"VNMHcNm.ުdwȄla ƒ3.u&Ss0ڑ*1E;Gь6,gl}<ᔟGf;ΗAuǠɓ7gUKسR#/СZfJ"F?_H_h־t{Tf\x&weŸb'HK6,7.fk R&.6R4 T$b<#A˭.O 2𗕾phf2Jv@l"-Db'ltvFZW8xAM.( sڬ3IFD]3Թ9fΐUzx'o]}d$F?/^vwF6Dqd;Ί)P6R/FŃYD~>:7 =2!L}tt5,Y:.]8,1j$wÕ?zG!^A*OZd7 C8~8GLazIEGD§G>2`rRLV:@o'^ # o11še%1x4z؂1j yF) l(N8ȀH A^dd, P1 .+-SX>G> PCl}ù ;*rr }OWLXEGۃNiڋ!{3Ae27>yǦeտ>k}B/#UeDv)q5!jzYJI"rP1@aSWE*w~GPxFp_]{'f0@(wiQ8CU% pi2;[wJo`^1Mpr\Nv˖k$~Envw}) VJ[Rɏ92TǸ3!5gq3ZkdHs(rM9 ­**BX :%xvM&ǩXfAbC!2@ÜL|Y UPƒCNYZ;TzR2W T |6ϲ7Xf.ٷסWD1қ[NfDB)a,YK` >ڷ7"Np3.{v?8Øq豐qBVXS*5:3BΘ*@Md"3{đg$һ(wrF7M~I潔;>|I fՎx" Nto>;5qew95seSlו [q=CC7`COw.e8qLvHg5ˣ3r g!H?n'tޒaJ~שiKY`Ea1}1jk1z+5Eǿՠ6Vn:6VJPn'k1q<3͜ts)NR8bͩ(,RX\"pj 7!5UnΌ@ fٷiJ sdJa @^-{ލԛoו @sQ'߲ R28)s| 'ӛHJyy5B--n=8")>m32 ?OVn8Mㅷ8'mV-::i /C{_n58FߜS9  PcSk1wACZ`O:'is}Tʑ0?bK5pyCg`pBrM-G9`VCA'R Pވ d*֘Ǘ+qYg0aű@Hk d.cLVfbb-q c~ļ`KzC4Lş获g$4S-ȏt2woFqh;ΟQ@;Vtu*+]ϊ x5-b_3bné?ԡWi{u/]'ssd ϸ[ι1 ;cV1aKEԜ\k)r9 @:8 +hM3w_p{8Z>yNuA_ţ[O e=#zy.B[qvtM܈M1Иy'{aŴ^# #,BD+bbzOQfE9O95?\z,_ ѳY ob" (dlX?{WǍ_d,Ļ8l/|'+K4f~3~#ÀeM|X".j}rAN$1 ~U(KFY, "h^v`}݄e//h2e[JbPzv"{oQYarRA#+b}lz|pdEmKJnQVpy[JqɻA7=Fv#;с\ DJN*b[.l *J>U HHn&o\m34%(ZeyU<=l2/~) }J?Z}yKt֌8ra#ԏMynl`?1QPy/-`#壿 !sȓ{EmW>'>X!F<`nVͫ&QPNSlwx_˧x>qF])!>x;"Ҫ[$=M {]pIr :P/> Iaz4/"5,ꦖK˯}roWW_^V߭PZY_kLUno Iw~D RYOͫ\|O7o_Ifj0<Of?ͫYY-_^h{? *-Z}<JAuf{R 5BZjv_ﯮj`{ګD&p=P1/7}:IڑbRΗ`VNkHf낕ĪKFbTנi>.3-03RԘV^""UeeGľp4l$Gf59$> +٤XPI%i) G( ɣ5)6M4T gjrGH=[5zbg6EUC*E5TLBk)S\=A;M7\՛ EY+A%ԄJ+*!ag' f"ȩv0 ՇP=f_=F6&D*+M/yKB#ٙ2\!9A}mdFcNb *|<źA%VuFks{tUkBE-192\E%zr^$`'ҳ{I2iv׉cdԥR{kT)eM+T أ*ެֿ wV$c-S/IVR454AD2È:k|?IC~m7aͼ1=P+'hF7G^ڟ:f2*/?世xã5?q/?,<W1[E';HP/*N.AQn=6 Yx87a5ݙww9w[?~Ph.EHG g!5uZ9-U8mu~Qؕth:͔A+#a%_'α PsȌ_ܗdC9 RI @9F9#>tGm0[UVҠ2U&g]d QL@( e-U1.IEhjaE?E0ɡs@X%BBXtؾ)W$~AژtИz $#)9:nm ˜2uĀho7ۙM:.c 3jtiKU:[ MvL}Cr&{VW 6_%lFDDՄ.*VtRT+gX$N4462hx 'b|$EJ2&LB:+Cczj0 N5NST?#v6hqӡ1ɜb..˭gQL t@OW ;:ֹ[-쫔Z,-nWHTMF7n8ny"~x*ku^?Uz<^[̓F6!9V0W`ӃoB/wtAn:v@t'ǺvS]s` X*ULkzW˵<[\U_ݗnXNk=P!d,Nc:s5-+ FtԁUa`ctWmԡ4{,s, D"(Ke2Vmؽs9INo_-ɾU_U vV}15~^MNW.?lU.o^ѼU \ԃM:]H~ԛ@|""HD8d"r۠1PHwx:R(]Q繑pdAr|-DS$ u\FxEGwԊ;:I҃^+1]ĠF_bZ{y ݭƁrΒnWbZqQ-j c8&3*z,z U{@zd [=r5l)VYวQ#te!ɍQ.FwEuCWxG<4P+inYL; M7KmfSF }\n趁6=iDQE>)Cn<4kxgțQ!'m\nuv XϮ0yIc< ja`t&$?I;CۇFJxSYq7Ho#0e W~#Ҹ"}Rߍ*[H!1DOsKtD] F7KYvtJ,BiC`bQCA;8=-ll;Z<'>=@ij1՞GG۞>.B'fm3̼>-B4SDFύ.-9Kb!{Jka}n4sx(!F9m{бjN'szP4Tw4α<(&?[b0L$Ir8̒&Of4ivVH;ѹ?,5Of𮛜0 pMc{&㥎`o5;& .,/bl't wyU Ƨ 4C'̓_~׷ZSVI5x4ďOS;ـ{+[=l4r{Zi𡭕?@+,<<;(Շ j֧8ľvG^n ;F%N 'G_<>ue1- Gd7]n{0hK ;=tl~ztYF_a%JiMNnǜ>U ;!yKʹ՝ sAНOd*f!LI;g+ӮNF|G_;}| 1k!"}\<7Y@D,!#)bͲNe„vjOfIf(V>CTcXE;!HQv2ڢ+B%(TE_bkyNE9G(/Ȓ#im:D}UJׂ%`yfv9 H gK)lErV΀ЍYjt7E'BKFFISF\ W?w雫Q?0 Z0V@|+DZ:gVGw%0tbB@_W6(RE_eʲ+>*4C+A=6TZdO Zrt9͚aPhK˝Ȓ)$ѭM#u5ɚMV(sszC'~wm|nTjSIy+#G eKm7أiL&ZTyKDM4Ʀ*6 -&[,!F6u*Eo-~7л7n16Em8''NiO1ǖ:U+%< 9!~$ALǩV?vnEO"!CD%AcMVDÇ\!S{?\[y =9`\ oFSaU6n ,ZuK7ff-p5?9gϷ7v>P6!HRuJ]^e03! Sp$w K;}US߭nREZ ǎ8j_V;-Wk2dU"'uL΄)!gU9&>%+@_7 -`?۫\*`C~+S&3azZ؍:?f>ix}!.G-wSGz Vۓ SB[JҀ|J@R%Sʼ 2' /Ӣi<%;v۾ 4T=Ww713; 4RiԢ*+NXM VSY!EU#YWa)>nZ6&p'h%٠}v6aUM\ ||8S Λ<=<-1CAf'?0z mH|Ndp?;fh dϝ0S3: kaEl(O,:CD ˦_"|q_g'm EG'S[OE ;pg{v1Rqgw(Bq{߻NhACzֻ[WO z'+k#l'p,+E5{ѽe 5+, ؙ\c6v{[}"C3ǘa m12LqV!+, 3&-Bdui8Ja(! F@@Ж=nt9ٮ 1CZ >h-E@Q8֚!Ky\cTWLVFH5$AdU$i`k-h"XYR]@h=:C:{E_?|[ pf|KΑkgٺ~}`)P!IO'0u%h_[LK[*ظ^i8 B5!7~qGwNW: 2PzuT%YNQ=Mek ‰ZC3A9^UZ0|p&Wsf:,D8#GqxWblDT !%Ol$MNb,fW7x6Lj4('XV xWY(XBKg0])%#! Bt#d#=\gH3"zǻk}Ahc̒*Svq#aC){0WhLڌBJ>>ڻ޶л ?ʦ=U0eu]TM]4J9#ho}MChȡ# N^b@?MX:G(*Hg.z"ƥD0}.L ØC|9*vw={u]{p#*boF^CdE!%;(Y1(0^G{ ;䗋>͚I;վnz6N|$t?JYp]e](BbNLJO*-~篥 yûe1u<-M*kNDQ3" Z\+m8b/w> ):(M/u> :z.wRVXV89щ)Vu!scQY=yC|tf'vdɞ PS=AFS22nP j0zS|\pE[Z\[Nrqؖ.D`5e VRE1Vreii)‚ :V2#`BXJnsp*P'*Gu]|(/ -lyKajǓtC zI@gMz~W!dPVW25i)6FBڃ` 9[0Bt66 ؘ싧@K6%! ]Ls{:@_7V_-*G%mJ |9\\@HZ)U6" X$ȍ%F:@jRb)xj F՜ha-ƣ՞ V al A+v l%ђU%Ip|ELfح*EaWKbC C,R+6 Q`Z!C2\1ָT`emjBa-ra1GXX͐Ҕ;Cc]Ռ`+*jc (XëƌƵA 0U !vCR]ga-k5$ҥb1ֹպ"I9LY8)a@VrQeil|} .fgˋ^.BFq}y #cM,7ubdT=We%BĒJW7G(Wʥ&{WnͷLcK,gӓaM?׏xUGcgyΪv晹FsaWL~g.Kzhw ;A̵X }u{gJ @OOnJΕD!U-r_Ӷ*D1H&a ##nX}=m·m\*P`uN8I^nfW2`B(Ri&$cq-q.Mn.r.c8Vpt@#&<\ARE Ějj!CjURCU*jZ lSDd]$S&cצ~t}֝;Ab܏nW;%c0m8YÙ -$8|{,zC,Gx𙔔!v9B\p ueT -/u 8Ǹ}S;\j`2iem͊0:G (m/b xG%L[ "Nz`|a}ӛC1C(yAo#PF{A@fOv%ozɕ.ŕ]D 4gyI |}*{l8u' q.4p?>ڻ޶з 9G m6n5_WMA)"dq lن =2hd3->ϱ7Fݱ ƇLbpLY1YQ$+|Y;/;?Z\6p(UD&o淶p Dɪb>pA;Yng7z}-(+Øg`vRJVd6Eзe_%|!uD/ :?rf$=ga YTX,#NJ$"Sۑ-4DlI (*2%6z]~k-L^:+f1o)@/* "~t ˕;H#JĪ! 4is:$Ux`a!P@< s@"E3uBjNx%ȇX]Q#u_HZh'UGLV )&r]1K쒋 L˜{`P [-@kJ::zH(|eHk1b.5 6pb.=:ʀ[v%7^&5xDd֚NUY)EDiL5k7$62އ߬"ܷyT[KT}9\16y3JqiaPE4 Lơ:7N3Nlf)cP,\ ]e8nzRm!.wkVVj׈#]0fc?M?&? j2epyE! BiI^4RMB@ ׈LQ!܎&E xZEE%RmTKOV~Y_jB9BM5&<=Vcni't5EukƧL]uO8A3)'== _Eȁ}aK;:uǻh2j,s z3T {Kꨔ ӂ)IT@$w_ݹ߯޹UΉQNG)h ~qרZ73Yr0!$էsɴ֯BM2NjZe_ rF`?4Sͽ݇"ȸ3ʰ^| # 4f?}uʰ}O{ِ; ny3σNT)u?ħ&%ďcXV=;& 7$r3)Y|zy"yåmmbϡ\*[FoMvGovqO'a|^λ˂__ VQ11%-cZzjzx|=I~ʑ6zܪ>o0x=Tէg/Zb|צ,ُ1#UL oe7clu`9AmjyKLL&3AHӞ sDDp8FAZl:⨾^#|SDaۘbb` ց&_ aU Ԝ  L⌃$)+^m[g7> &K̿P6!&b\Lp s1ǘDo]<d;\R8o$1.zJa攺A̒ML:bɗ[e_?@2 ~j2ǔ!0*ݚƯ|lƗWZEޖƯ̿+ <9"ͮ՘ₐN]דDaB,o1ElqUvxVOF;Jj$񂊡hSRaG.+Y ,xg>ŅObUhYQ 6{ mCA,k5ؾ#R1UGbxTLP{EG 4.{G v:29)f hhLG륎#I$6R)aRN53jDŽQk"Dj<!ou[ HOU-BFNPFT҃ĸ-HBi` 6T FEK(SC#ޥ2YmmBd[5L95g՜[rAx,lz@7+5Sl4X@PtIY16iJdT)a+]6Xma tx`Elz>\j^@>s>!cjYI*~Q Hd /U }bȊs͇&A_oBӇd!yy8#˴a=Ξǟ?)6}G]틻X ̞~7+u xJ&SJD//4V~M%IB /$.A5 dN+q.d[v}غY-_e±yp6+K͠E@zqk洳"B9Oђ"qJ/⼠2>|[T@4.}ҿ[eukj ~j]\=A }a6QzgOu "ٵKfp,^tRiSd2n#߅ė SKW=YoDklNc GnN]ۈ:]Dhޭ>M-n}XoDlJcu%}ץF='}SˆEl5'?x-: ZQuUVϱTHq b^U%y>~u0>no:(:yNceP/qI/m 07hUG=3pgӥȾGsmf\{:E`^ C󮗌m>͢0ߗuQo-4.z_;s|6=Pڿ x5|Trۙm c*DP X P"ZŲ6e^$pM.%}gQ$?_E767Fr՚7yS}@Pp6;|ހqs]EC# W6Rh Plb ֝ 6s~RpDCWic[SPCN{Sw$txvӰHT:p9 8c y(h 7tJS.žnRg6"qr>C9ߊV;sHZghϮy? b[}A|N$(03符y5iunԭ҆#mE8Jemm:ulcmk3۠Q(B a\J6J/ e p )"e)f٦vÍWlQcJ"ցbƯ!&)+hE:S26Q⊤Pd63}De| FPGITJA'#^Hf`1 Uv7w{KkU i9D!ߟeQ QQ SUbj"Pzo\_fr^-. vvvD FcÙ6B@BUub6yD3GT+㭊rbV(A8~CMlԡ2::o {Sy{@\eg@\2Rٻ`X'7g+߄4%@T޷ 9TҭlI̓D,V^%TyGQFŸN04%V扎*OE*R5כcevPteaJkBqn (N&TJۿ\d"SDtQ,[8*lS̸պ;KgYxڟ%aB:{n{i@…Vz jK@[){ qb1LkډNk'_UhHGrU]o(ؾKvZvrBNng?ħR/R(.u_}[sQv~BRŇ8l˻/C'Qم+~6\Aǰ?|VBF:"}#往t;k2:4pNYc5㍓bY }iTl:тy6exrwK4t5)t(St'*Fh>>=}icfZ㞼B G<Ml8ۇָv?Iϫ5+te˜iXy}֍N{XXv9]@8їTUtnPOq~Nu|(Aj7EUֻ*컷9:ڏ:e}}K@&d7ITߥgY BwR9߉o((]l8*ӆ"pJJ 6fMy?./pOqlvd^˚n *opvS[h o~U[ow2 UZTQdD9%O7w:K$OՊms1vP]^Q㐨Fi!NiVIk:˸)Z ϼ_( 蠈u^b>J`uDWxN=sZQx<k}-)©=7̍@+TbcPO&. Cuit)M +$uE޸fn?x)8.:z{ <}v{=FCb7ys[2_fyu|c0O?~9j5gB ]U"m@eLycS^S]1΢ :2xey ZB(j߻U+ X_-xM3w݂gVH]r6(nr{R* h paؚjwjZ^z8hv9^$) nnRjvx Q_Y(Lay@b:P.m^fnkQ)NqSah.F)IvbbW (KzmhK ax7nj'e3R&51 %ݎo/yYWSR>Z/ܾ2~u%G ?7\ w/FLǮbd_h=2GurzpxuAB+3J(g2\bW!\,G V NNg[}f"Aڳ^*RM &tHNxˣ[hLKRnqܙb90^t-g `'&9ZHl),jz ƦqE\A j 2$ )Jo|C~XFP՚w)!ѶNsX6REr?ᅯL9ԕCrb,FőM/bҺ. \`^K]a*@IHO~}ưN~(^_m,WKZ/^^RK褪;:>`1Tk:1A-xi: ga c*$R] DccLMȻc4]D2B1*:A"4rǫִ,bz64U+,?;8m&)y<&}SF\G 3 Te;Npcl0ikfR0r J.?JX#< 4V. v\'Вpi̒3}r|x]Wbf'r\,[4RdzZ_ϷZYKop& T#K[?nn,(@ީ̄K: ݡt*Ř$ܥh&lHNiDIOYgQn4_Rįs57k-zZ;8ֈ#ugqb<<ۈ5ós #p. !`s}nǡfeR /ޜ=NPpV³b x%)X [t͉sԸ 9ܔɠ$*!`F XVxŰytƠGg۟ (nY+y?YLHVhTwsj:]1%H{(\B%gc*csQO"\1pT ( "cXZJJ*c}oI%龜 K+4iiд^L/.{_xu4 S Z0i8B #Ctфg1Cj i>?ϔ+FnoAxWZJz5J$#=XF? @n6c e|q4Ń.RQLД1D-b[2cV+WH_#a+9` Æh(aQ9dCip-&ܯW6=5<+d \ҀiZjdU*cc琦"9o=:Xs)s^?ig{~H˞t,X.9?x#:da(Z I[v⬫۳^[Z.=_M`yoD][?0DWλ^W- 4[>>kt0}o%# 3xߣ;V &e뼙xDRZWg?ldvT2'Q`Cp1H3ƙd: -Bc7 9[@FVl.I*yrҸp&w1i^R>z#6j@3 \ol4Klyi(1N!YIS.$U@{=r^PSy+BZΐcW]NAo=,7=dپZ(''}H}a"t_2K JWP:tc_/vӠK;4;țX SGK;Ywv}h_Yj_=ՏSh]݄=KS&"=Ρ[i.ހCp!P}'˶?;;QiIګϰj[HUj5*B귐jynʖYĤ5»Y{R8[e $754Cn׵@Sr#r31ve_<ⶼǓcby v%)9eBGSEw*ut K+1ď7>F%]CgwpUzW#"ѡ(nX]L`Ea\( 8-.]zUH'.MX[%TsvAwJ3 T1,; ԛH !FJhbXo+k4 +e}y:00[z2 HTm# .Qc*]#&GweL^yj3LnAU, L`3t鑿8LA~zNX=~{o_WƻG`WVN_\w;[(//Ǔ/[\!j_7fzQ'x>,ףƄ b 笈/}@q;NJ*MK6:sUS7,Ir切c'\X[?X` FFB@HKA`\bÊPK yԟ=Qhո~͂8'$k# =H'rJĐo=Obm<ػmPз1~?Ey I1;S<mAF$:LJ6X< !J(K<#:M{Ts&9~K ]\^ \wSY[0*㔡$(Ί]m/noܼz[:x=STu`/=͢b ~x^%oTnAsӄ^0}w+1GNmR0?;=w;a{7$-̥'>Ujů y"Z$SSSk0o&t<ڻӼf)Ũ%:Fs{ҺHAҥKhb61]6$䙋2$NJq:3"ӛ`ߌoo> QM:/ěѳgtZؾ\ 2^b+ݴjQ]m|$T0(Q2Ju{=i4T D+ytYY֭8rw]#4֧zj4(/a:XdTPIA 0c`9$AGbEwcZM9$;kF_, S1Eĸ[E'~Ϫv伣=^䒢:Yhv4B-_co1/Fp0ȰGop-Ο[d$5;ܲ~W@=zGatvW9Je4e|u>z2[$Adb_<Ɂݗkj󰳇/ h8&(F,h'ThlQp ΀PT$ǒHz9vqlESP<1t49ؚW;BgW4Z oJɬka3k']xJ&'0fN8}3'5CRv؋gW29.%V(svp@I%$HDxODt@NY'ٻcWy9Hrŀs$Cv!ͱjFI]'8TS %RMR3a{FTꪾnAhWcȐch!K'l;Z䲫&?waq;vM(7 z<aAYͶy\m;eWc~[!^OxuvF]LkX]B4U8i:I8A4jHFi\mkت0aU(°S[=n^N.o9l>ϧ}dtLkfeޕn&Mnq!V&Z[0Hf2˟ys9V{gsj7%E&L%Zpi i4C;AT# 4"9cޭŠj?׏g1 -2tUNp1A-s[9 ӹcXpffZހ;IW. UZDFEg"8Qx8A}Xm&+P1Nd rBuZrk.#E65Ge>[e:H۬v] <sWnJm"egfB[ӷJ>>=O͈3>o  >@>FbB-*o᣿o|{CcD/ã? |Zoǧ|B+D #OC08 ͓۱s=)b 8xW7-߱rW %A(*nF&kn-oT|!š{ҁF'(7lC2|L\"܈.\~:ܨ] ̛T =뷉A=ddWwcX TWYZDUbrnae3A'}CtXO š Rvej?"Z(;/AA@ Rpgcv @B 9ONkZ"A fUAڦfb8*<*FOG[g7,^xWOt6K;\U=%CIYq<>Ma_c;w_'sڽ-oڗ(&&JOwd[7`pݯd齻߬U^%Ύ}{%>/ܕNp0w!'+RR`UuO7Ҫ>Qa(V=Ph.wv)B])QSAhd;QDP 'P[PWMGT÷.ԕ)6aSƗ uX6v17 g`{s1Y^C0H"Ej0N[v{50;E]8퓤ž ,gEX*pl06ž %ѦhsZUkPesg©P?;%+ɟLg<xQ\܏5~Tͤ( ҝɜeƏ_~QEo?xi./'ZW'zou@Xƌ|xi r}zsNQ{>רǝ sDmB,L5)EYB¼$b.&EJVS UW+EE*פu ~\G3Guv_+#*hR\YWgJigy\THn*Tgð*x [ѐ(V'(&k@ᱼp^ -C:w(#[W]CѮ׉0.dFB(_2z| _eh{VXc|°*Sͅ.?Q4a#wS8MCUa \LO7Pu/c;[[hwREos.A0WcАwHmX?bB4b"um4$84P@:X;$V/:ɅvlĜ!?C4^W,C >jzEwaZnb[&]aĒ0[ o5H,C6mqn4:zk7/f c! ;'k%W<4 +4 zէhv=>OcTK[dfFeuB=J}yw}yLKVJEMG[א葑 ^0X.FKlwuOͣpFv+ z856"-RuEk/;TjqޕVAM{ѓOG"? k^ 8C"ChMBr+L2ǘ4`2E,Bd5w/!e7QJC,SPf$Ju2`ȥ&q"& T6# 9e\=E0[$gy=[j!QyO 1׈)0 !o&P TPDf *UpʜK2$MS1H)@,5LАB3A1yj:[D.oW3JF  B_P l'zwAhN)g4˽ΕtHk9['G{_)HTMjѬ1R}8rr@i3!Y* Hx}@"&!Z#@`&!3M =B2~CXpCyso#= && ZBoY{&"xX9] \N CtpRC֜BDfsS4 Qc%ŠaE/UOm@mL" \3oj.5aP)TJͥfJ BŢR(ɹ;{OS48Dc)ZZ`cg(h%{hyp~oo֞ *oM,U9(QL҂BD)-1Ԭe҈I?Rɦ3ڛ&8|эځЭD(X?V!Wvk&J pūG5XjY63y>A{TL[:VrH2Пe2oASp2jڗc%q3ҋ`}93Ѣ9m[ (1*0^/Ɔec3lڍ 0A}k!H ~:]GTN _Q)#-5]<8KvU9jHʉBrLLOwiT ]Ԙ) j3VY9MxQqEHӥ[o" Q"}vH7<yuaO)F/:Xp~%E 8GjkAxV,B+`& z@?X:TÄIjG׫)Dogvq,@ Y2eltjm8C܌r IVoF j>3n8u@oOs&8gS Nঅnތs~1$;e;Ce4{&g!o![c8_+AS4+AS4-E裹\Pz"z͋^PZqkJVI;5هZƣuG2]=կ-q]}rkRL̃!ԸmYMPۦ| !fD\J(g]$L k $Ew̐ER C>9)/ntOS !%`b)}ȗy1Yz{s9Ygշ~41Jʖ;vTCZ6i}׽ܯ.BXR]JiG!][L&Sr 'r`%`%a>.TVՅpB 7"{'G/z33OCl 74O_gO~}@HrMCfAU$I%RrG1ER$ fr $ wHZ0+]Zu DKMɫB Sm₦)iSĔF'%])L-JHM mvBxb I>BK')˶a̔Qqblj,I0Fki8gb1P"51MfX aK9rFJpppþB$bD nZaevXB5.'lO=Ӄ(!u5A&s˱5Oz%OPsa X.qpZhI<(qj' iÑn==u0Ԅ\׬ADd&KFV4EF; % ;=d Y5x!8޿w|zXt77շ|Y^0KFwĸfv2;tx6go1[bP-rlWʱ&ĺpX/# ZDVT՟:,_]Hۿ\n `k ] U* J ̌W" xxEaWYvĒҊZЊ\LVG*a{M4ֈ(PhEaG.Ga_KQUQxk9e˧;^RZi_ZCdYne$2,t +kP/|9pV،b)BCTBG&TFnY 94``E3ogQoP.gxA-¹L +Sm~-A A|쾃DAP3!u;V@r|Pv\ *@f=YUk5vD!fowe؊C ]w"ZҚ Li/OlyF,DYާڤ,Oռ}!!{|63NH:GT$ӹn|2TNV㓡RpXpΜ^k~J7 `ξzxCRofkXg{@ZduKFP^}_{0!Et |tۣz~T<-*o}7nDlY'5vqo4i?mc0孟/5g) Wd= ec~ǂIk%VG{¢CvB{[!EpKr*SQb†u3ШxD떋A䶑bݎLM-?)Һ!g9:E >شnJ!X\ bT'vH18o7떟iА3W,:^ۻi`B5Ѻb:mXe݆Lެ[~ukCC\E3toϿ07d.,%9'&oQ'+] [P ?.'Y[w8YlǺNٱDzEHF IH[*(HcK*k@bT]\eD&Q0]"ƍcY3K BC mu_4YMi0HY2c%S#;݀{1` #YU]1S)=5Q`Vcr _ V(-6ʳ$EѢWJ5|M86@ 0%c 3r"]UDXI"WWV+Ak  \{|j%g7KRH#"a9#QT1[Rc\H+ R49EiPߝ8W^ƚFbM'm.mb՗Be0_"<܍n>(NvW?zu3M .٫eW^+ɻRQ\%ۧ:oǏ[D%fvփC/gs_Yh4P!:K87DB9% x(K}IJd) HT,5j5,Ap 理 E{ԾS!UK#uz'V-SDde Qr}00Bc(" Zh9#GJL#w6rJakEwu I+[?}u?1%^|F(kU턽.Fѽ}1:t\XnƬDWp(/ո::(*i.ɋq3Qi#!er(ylMN9aHBG*t%mD D΢ݙ1',!էv@:AldH4=q]aS'멌eI^*Kv\(reH#*);*_6w$tG^tA}/l/]L6sxd@ JZ^/h40uTv33:@hg90'_pHJ0~_9: .ݬ^'^V[dFP]mqeSsSZ mVj-nz$QZ|7УcD 10!^ Ҷt ;FJӼK+DLw%7"aтJrRĔ-@t/B9*.`I3-)ҖNBXf+&"jVG O =NdGҌ115ঐUh@UZ)*#mU )$U6pm65>c)Vv2$A;alԧɰ|zر=Qm,<00-RU[48%5؍ruEk89C̩Ե!9EXcJo_VQ̉Ftf]$ +3w"C"Da AdROŢA#d"ĖAbxoh'!70Ɛ!|(hQoSa ^Q^^"ZCinA<+OòrH V4GJ2j#%EK`0~_zaϽ>-B^mDnpgԳ.GLt)1]6f>~4Y3.uU^u Wc\ 5L` ߒDY[TS/%@EP b3i:AZj?NC.ug_5$ܟրc(;cZ4BXZ''1 |ICm)1k?P Ig7# c GR(ݿ(r{~O6Uq'tCɓ 0Jmljz>KJ{` { La)fL2 )mzE՘ Ƽ?Rp(XjoUwcSiۗ%'?Ύ6 2=2%rNXB~yN;%j%;) >h('-8%eF%+럒 Cf~4[dL\)CN툗żaWL3>m87Dw"B͇@\@Zyqa<7 H(E%ڒpI6/~rEJ}u^q4?>1޺cb`(VgKʎ.F\5|!"& BK!tMd-+JE)!8r*()3ʊBJUjH#Dlk(.{xJSbw5!"v\G7IN gJ3 )p2*pJK(gm$jғ0+aSN;tYX9a)i)IdAԊp)CJֲ*sE6hJNʢO1z_L"xzAUq}]Z*tzꕐtd:w9W Fl}D-*Ob`?zHBl/(%lq  T;8jY;Q z2eG]ӪEk#N{ :T[DI:tB Ci5u[э7r*(AC&ҎYohTU%m k.i0]*|wpw>/h ām|zT>,H׫仿7l"Ɓ-~WnJAW|3g lݛ+t:䘰ou_~jTMgC1۾1w7lGK '&JIW3mGS}5]b>OMc4+k?hdW?c;R pũfґQb DRX]A9:|2Z ݴ-μC3DM֔ʍy&Ig8ny'@娷ٻƍ$WzqB]WfU)&;vq(pPLjT*d(ȶC%H|Wee!5<#؀8RDz5Zdj}&6*4tn-: ذ)YBN5Tlho0nHs^>փ@SNL&EK$$D1n֔pc(4eKi WBwݶϷ@ۗ'}dbSRoꭉ@G!猵8 Q/wUiExˡa+@`cWc^s;xNFu8HQ:w,qw, 9l8JO 2K;4K 5HXx<.vBnq4x2VȢ^;zOQ=`hZ1@p<z8C(pO%T6|̶(ƭA0)i-Lz|oJ}|>W޾%ՂVSٿX ]~/*:BDIH\8CUA|Y^o_彾<[[uvzWɧo+BiOoZZ{7)'G9rOdKjFC8ЭCT0)tmW0,ѹð1(O]k7?YGntǦ58 Ixsn'#[ Ҭ>zmuAB+Y }*K@K 1h4 `8PJRA B\R GjI@qιS#d"1PګLSE-T<}OUr?IRL &q=孉ϳ\nIڇIQ4 KI=GL-Os(.ʏpZ ֶC82,5kzJ &7%>s?c1dzq{yl]ۋLJpJ-oGw?(L4I6W. \\IeMﱬr_~eԍY"UybYKdՓ]hٜsT1EW͛\ϥg.J.^^풕 R$ELq ڭ)>u6OKn{jr"D[we i1}xƆthw[Q[ 9qm$S({ nڭ)>uܖZvݚn!$E`@)ٗjkE;ǀR 8oCR8}l\\ʊȝU(J3xCA}V{GKV58`%r{MKfU6IWr%]M0:0]@J"}=PIm*ˁB@X.DKzӁUv"(;;|uFőF02tf_1$mkZfk\n `AP~@J/x-l_Րڰ >vXY띂{47+⸥^ho8T{_ZWI?B=iћ4k3^o2T8tr KqߌQ?ӡxK|iѕMs45&N7tHZ wֶpК6_P ~kZ^]\,`p!Vhr9 ^DAx*.)^UbӶ3:c}O}ӖqrbL99e$ϕF"Th Zd@H*8 IMR);pe7!]òֲ}RoW[Z/,@{+2*wSdQb3[g5y[nM^ c$,U' e9WF$)KܑLjDGXHj!(NP5hqFv-fIq+!,oF^,w+QE/.O^<}2dۆw2OYdZ\K8kgRK;(`_oO N۾؉Sa7ܴ̚٧o>Ǐo&oy5ywW "#PpC9xRLX <;oǚsZ'ts[46܊Wf R ʄ4ZŚ&)Ró)ּZ Z@@}S32 <7H$N1S,'JQ)4ÍV{Qff:]%Z-kPkhZw:ya>]]z>߮;oeO o+ǿٷH_0-.΂uf}a?o]Xq:[,&1Qg&tbܬg(twgFe Ɖ;?Q%W@)T]c/㖋jHM=DA-$T2V[#CIe57hɐr%m sZ!E'] 돖uAuia|Kա0iz(ҥ r`P.=)̮(2,?5VgeqI:unՓ2Fp$ybZ.c4+27{?b{' zq d NBZJv[HʓhUPk1ٸ]rBh Bn!R{R1%*[^eg4)nj H@*DYn V1RgLjkR W-Isd|4ᮍR52պ[eנ\q[Ws[/( mٱ*!gΗg nDHh++)7 *Tj#4RpA5`CO?bzN z E%%\kzwNFdKAa[_Y\ꦃH-Uul 8jw\r< یm{󦝗=񞯵`hBxM㝔X<[N=6QC|»ev' b8 y6CC!kvyR6ky0kɭ|u5weR:0$+mvژI8[+:A!bB! abԐv4k3k:U*"HpG ύUV9'.vbU*Wwxŷ/8$U@'k4O#]?*Š{ U\P"O7Ϙn d3#bI^քS B298c\Z/giA%w)ecR括iH(|g-iAi*G4O@@Jq%^qy5B2BԄJIUP*ҕL- -{N+aGڻ'Ͱ :m>{쵪v #/F=C[&@}g.A-Ğ|z#Fǻt^$FIhY\jÆ꼫51jBNP냔RRHhtTNuiݠ*\3 {h@:ڨ)'k՘s"@\B9<9M RCZ]恀"JqY( CWdQ vk\VTֳ͎u2 ȕTW1H ;Rd\ `$[ReJaW-"!QNIt1 jZR >؝* \]NRo<ٱ+a; >*W|IL/TcQKHczkff:(umіm@ j(e^h cU6-6+# 4(Q&H˂I0NJBX)l'$R CHD $K)bIe,憫4-Qɐ\ڐн)eAgڞ+k/ӌ"i ĝ2iUc)%.\zIr9e5@6~["]޾زo_6NTrK>#vxS1Tc0Vxu|hj/l6Vj$=}=^x/e+1ˮi<ݩiXm_|K ݛ_t8?nkS7 _7 7% D́$*a̳Ņ[2Ln& sZi B)R}H[q+O95v١uEkΩZ|,?a>>xyrz}_d$]e=u^5ToDu9 k;^PXUq4 MnPlX-ahS6{uq`m!RXz^$J3͎:dmBTKme ZO>8cD|8jIQK:Ը`F΃ێ2dOێG׮]rQyۑYLJXQcP39W]ecB)D(I^v;f)z3;c;(6w(CR}DM)H*He;lI,_yypMs_uW"կ_7iry+Yt FP`8v4N1tD񵁂[-ӔbOu5A/we9"ٸ^Iݲ&Fd;o8suI`,rjQ0fl 7 Zj$Ñ231|=ڠmq+@-̈́iqP<.S,I(Q{T 2JaMi\I| ,"oI;4ЖOh׬$K{fo-Z 3>C>Dx{ z:/9~XGr!)rۃ?QNȥbsF^mAqTR¨[ї5' a c$VH=yF$7&7ɍU@!ؤWq4*+È"X$AR?< _<}Qy3\S(4FtNy{,~4VǥL(W]73VqS2I)4g&mXZj+jM2jH! TEQ56d@ ;'{h fccu$4bC8M\i,ר2 LYyojOʈm%=Ԣi =dU^_@$XJkAPC ܧ@* B^ȉ28ĖTdZ " Ѻ5kUsx8mx8Jkihl.d!pa"%pzI ( ~IW0(.6%DJ!2)9j,Z4I4)%C0DKDF 94ymͣ`4$teTn ;, 56`s?SF8^ΥwvY}qণݛlsPNвS,#'T!xV,̓ c1jqd"2ƤE,.8Q̓+ǽ'bAra4C1WsMg^rAاhLco[BH:Uw0x4^wԆ.f Ak&tZ %TJ1BsZ P[;(W32 3\Dm`b v+) csDDcYilDq;Gܙ(r&NYХC/NR@oo!DԐp)i9|1ӐN>0[ð^' K!Ġ$,%RjEhHܒqK1"Ch 4\b *7.)Xup:,0b{0jU~ܶMUPx@n2F%_ r GA T~/v|7ʸ}VɍG\iPf> .p}YgƢ` uRE$G x[pg56<0X.L6]}h_`17#w;'E{†2&;H8i7Ll SNKtz?9gh!=܏ܫ'.W>?ת rEs7L|ѻȅv S&VyAS޷{[o؇?]vc!848ɲz2b#*r$W{nEܽ<>";RYmѳ80-AHOzx|8p+a}S圥J(ee4nBx#(*YELrʙB)L:^S9EAi2;@5W9 Ex""T.9KH>H(%`xMu|(R[DƏu*tΚ;6w1dhKqsG;k5}C1QsClQ o,jXu$';:ޜ=P-XuxѸ3ᨢh﷨Vʳԓz$4-ʖ)f<( K_M9_'z~y:ӊ1f?[(BY  ӄޤ]? f޸;ӪOn.K.yJx~[/><8e'tHK6czK)9YüEVD֔;%R>H\ ,wn;bH"יI +`.6 BluU&ې]O))/ M.;,=En_C祒 a>[INx(S݅vV!UNJaO'~nEL SqU" eT6N̈S_3ROh- vhl;J;FGsҭЛY)ՕAh/)dtFkES4ǕP"}(vMH'KfI?XE(U,/Nɵ\4]O$= JG#ȅoECDe6pHoCzқT满bJ.+|Gt2:x!M 9s?|/~Cx3xd՞7e b#m/6L` [l_h/T RNWAOjA>^EzhFs2hBy#*i}cyQ~<_d&O|惟.R \Ez6o !/6x\v(˳~;"L q~YvTtE֕{bI_ָEP$w{S^ Mb{!yZկM { 1:ZgM9&L$ BAD5@QEw`Z7gj{ȑ_p7+k$Y=$Hf `s~[K ldVqb8.>.T=UGuZcEJ, !` ":IOХ R=G~!'#-ݡB-z @%S 5w@D[$![zt/ͫdp{XUGY^IO*wԷ,)7Y175R;v;/]A[镳q1ízӛno7wg^QCnMy;73DŽUyէmmåN+lJwPJ87RJ7(YS j&G)%r||ύ:2GƲF <,kmB ?5jv*.].]Qe"~Ͽ|a^8z$@p5\Uژ=B񑍴e)g`Yd2)02Qps^AQbގzP8 ciCc;ڀ>yiW=]Yۜv_?0 C,CQsv =J 7Uj ikhŚN_?oi$߳QAdbbz=Ao/: Fx{{Ilq7dMG/B6 Ԕџ?9緳p`4Ч|0zU]>r^'ɖdUE)ҀdF,Ywo㖛jn1^0h8Tx/ G(Z@YཉXLV4[4:RE?+ژ{iسhkΆ:_O@+>e^\yԠǭ Ϩ#wTû^15&{@kW!Hl- JYo$_ Ic`]X3 lgg)Ka>ddr}7_ io&v\ފ׉Rzrz"Tĭxp- ܖ>#\Y1y͟Pakܖ q8BM5QLl8 grcݷ/;)f4!Ev1ђ6Ƒp#ƌd<Ň+fW^(ٕոٵxs\.s5Re}mI3RJʊSw2sT6xz]M4 5]E7! eܘU3jZdh3ؾTaQ7 }{E`A A~!^#B~ʸ5@@Z[5|`"@3cf2w_(m([1TAZ,BQ4YP"tOj6CK=ѵڈ.uUr L,)E!$*VE Z| dmg !C{- 5ڠɂi';-p Mzi".Lw: jljk"RKFwIwblκu1d۲ӓ- 8ơTn>nvBWyFcO%}Ӆ>!Yqq-j3 U:d8{<]Lg|+b}@XKuG$hiuU2(Y)JOm!NLv vm;Cl1ܺ>`Ԩ]Ŕwp7 |=i`pĹY?U$ g:8Ǚbc2͇U=_hLh~,E +O=%zg򮨓/&lv/O%+'ӭEbw.H9Vm% S=_|k TG6Zsӝñt\Y>SY#lPUoL?yƈ!׌Q3GuBT{'?\k֫p񼐻UxV\Rk 43CUlj׆S]ĵFBǥXC, p,)-pSho ;t{EtgX_)6\QdpZq:  =Z Y\7\)Kp(&Iۣ5s1sZ 孁FpEP]$0_X#\{BDuBIό:F/-s(L@2X|NѡW[zd"%ISV'nͶHlaP>gKyJT,3!XG?_Fz a UL~ɤX5+>]]Q$6RiBx\YP@)77_m74Ű1$UϨ1K#р&$)׻MEUz&p@նx"MtF-vjNKZp0+e|_@Z4Uttc pb8WTjrjoJd0*{׻;Фe٘) #3z@cdǡaP}GЮS{56i滊~,OJį}e_WK"TXM%NGَR,v/X!{W)>їO8S$j 8=Eey֧ ZY̙T iBt/PV{iEJ5`LK˅E37 v9vmZ]j=*;׼O)( j$^'rP$0 ԭY12!?#(pbyK&  \쯑unhuZnI3grY?3l=.FWUʡi` ȾXx Thgf\*܋::LmSC0Y6Hp@.fEc霅8F7p8phP}uW*8A1 ֗L>`JŐ '̀cnVo+hCe.r-rd`Xt G00(QJJ( h-;>lTZkǑ[IB_fKb'Ocgia0T׼,7/(UR%\ ~A#Ȉ/RO&TL'jЇJ' Ll(qJ ԨdujH6*g9it+$ u T8l8i$]4k3&c祊cŔ*rm?Hiu]b-l,=>6֭߯QjxӮqL(tW3eTOJ,׹Cb*Cą1`T*=\l@l9=g_t'c{{Owe +SnF#B[\&[7WU.*s^\yu QהKcp1 3A?R>59蟬[{3L b.Uy1kѻ*'[?0o4Z}'VXJ3S4Np45ߕMB GhK%ZJ:pBSV9Rcv#:*,yC]a9dfhnQhrdrHUKK/P˵3eQ{r"pBuQats*!*hD@l̛RheԈX&!:E%]zg+ôgSz,K"Nq38ϴ$ApCuJ㜓4.}'4 Pd> Qk3^h &Kmm#3fhyYq0A.[ʜ)j5,F=) <9(~rڵ ~$ΰ6"?>{a_F pZi3` QXQB7'+Ik_߯[6%s^d(_Y{_ߖ2=,>-(ا9:ďO\.?]_S4TL{zDʖdL^7NW-n`dEWs#"peeߺy|i%c,&F^}>l,lfY!%~9 lf \Q&¾輙43hI*Yص{2iYqlCmLcbeGA kՌ5šV,&tdXLMAC\#19znFS-6.ۇ8p99g FDöJup JI 4o' q#qY8;se'a I:?uY]Y[oM-c:wr㫄.u/#-HC\9%/ : ;]8`5Vp})0TQ󍗛chV{g|}(&%1REGtq)/i!MN3"89siZr]D7ml3dU1Ė\[J^,q`G%tdZH2i( "?ŪO*sU3)?pefPUɕBMJ%pJB +JRDSƛu~Ejp2b9j,Zc#Z9)!hKMVRg<:hU9?>8o;YxmڣF9g2́Ia<=o/;LZBѸ/_>|w:5xqsQq&Ym 9zztY`J"X̙֨|9"!k{SMȺ~2i{z8u-F 5j(?^egKcӂus##3"o5F9~Zœ.[^͗ggA3:j3(I:&r:fPedC96OţmΆZ5Ks=W]φ^ RP}r H/A()Ӹ{RZ1EƿVȗ0eKcVvOLM#NA=q T q/>d< aeS# 6eQԍT*@pDtɅ:[Aڑ>/q}17Vm VѾr@$YLƍ>)&84j*4a*.c,I0Uj-\J6 }`(8a*厩hh84l Nd$hmXU aD_tW)jyjf@J ,*8{)rRy?jz{)Ass]n"* ("{&3 'Q]<'_V {Q,>] #XFUCHab oe_xX s>7f.ĘTk{K^Rg7Io={e(:bR_bQ/?-0lqwGv, |EYCC_C j@y=4!nN䠩w!WխG^_{TZ ~qbBu vه-+E}"aZ˾{n!CKPy48=dj0+5!UF ; 4!5or,Xۃ͒~7/7\}d7Ш1"1K@L)HI˵M}H/Mmyu0]S]ߐS0C^!NPPj5$Nn:vÆhjsb׊@#+Ё'gpDH+09oWW3HH/B9A+/r`]iPR/xh v \E)V:H"Q{lRj]?8M,ȎTw.IC i"BbeI2h8hh4Rq Vrd,l=qj8=S?J(Cyhc0zDOEg@Aс$ e n@:ѓT 1I5f#|ڳvUPܢkcn樾+-AX# AM((xin>jI3ܮ67-N cX, M`@l S2Z5*X~tJ*߭j$?s XM inDz$-yd"Z: ae4u\ VOF:l$T]@$pIpc RT̊(Fў6+I@o7wmyndkf,8%4ֶmj##P&0m;֭*fE; Whi"Ak>c)(Rj̣siVZKfz\.| nK4tÄ=(< K%R;& Y8֭XA_tz=N0!V5+}b- ;7{:T1T:<7\D2ILHȅr$ܢGU:%E/G7 aZpuuH;=,tP Zah2[W kԚ9ښ-Y-A{3[(}%fdI;nm^;=APxNx?0Egt: Am]X7-b{7BQh@`%30X:> ;ΙE/x9WI{|'1 V;ehӃA)x; WNSaڜU; ׽SKmKLI$j8˰>l[M⛽F|8s9Zz9Yp|cҐ+9tz[G6s^pƟp]ӇR_G|\w#㕹w&ObTDJ>5_};OwDyEL;o c堈{aܨA`Cph&8tF!ROZ=nЋkBA( }c5%DR2iz4|`>o9D||L#P9 TKrFۜ)k OrsNKnXi4I-)Q6E2fbTG:)sxNR+~+%*ʍ`Zj.MMaUMIV񴚁5ݦԂh-{!NLa$=Fj+72aYavqO~f/(yUoK{ĝi~S~{lqf6\+~iG /obq{p*Y=*dJSt)tXv*=(a]'u ;K;u /SI'xtu ~ruU;)7?οUAFOe$@N<,Pu)Qu? 3zsPCO.q{r.Rvqw=r?\ӧDU>t4*tRl0}w5*$Մ%}g" ? Bw% #6lUԡϊh-02 J}T$cn/#y?\ٜER򿧋9pą,wE&4;aOӛK}71{HOdJ݀UkqS?5Q!d)6!+ϾL}- w|wRދDq0_Sas ]|oW>=]w))u{*j.q^jGv>͇^wWuM\.SP R9PTs5PR|!Cჷn¡FʗH-؃sq2rlrK]!L(&Ջ)P齷5{.T ޶>Nt^I4mFCiO 02`q{Hd$=##;${jI۶$de{01,] 3=숬8&5ò$RI4(S6^HtUѹ.F4) #ش&ecN)@2F s$Ąh@F_Yϼ("\>5^>@>((8DPuQ4dO?jP9 ;nd[s @FB9v̩j36G"D_?@n1Y^[qGtTXg%F.'Q#wKoq)(ivM>.hl4/]E1G  /PN^9чۤ`QDc:q M=pi` 4 *פvT4 SSE~r*2 @[;Uג 9@jJ:u&~*]Kl5%hC,52Ehn6<1(6%.D䴟z|[;w r 9@jة6œX2x@~F8@fȿA8.Wz*@$bΥL|wf(&Sϡ4H6M9Jn|N~04 WLNQ( dL!thU 2>(72h;0h{ ^È 7ÆQ\D^W$T(E:h7H` =|6QYu<|R_-8)SAڂ9N5@BW.|uzr\.͢9nw4}8?'8zCF3ݣ,J >kے`ĔKBCC4u]Ս J+a|ECeR$$!9|sԞ_Ie?5UGT3&?GIZۄm͈9r4N|+C*˜>U9緘T2FNK_b+N@<+'=a\M ^x3iS/(/mUM۾~B4Of2p~S_D5u Df*ޢρ$#Ew˪ DUV6Pbr۵G[!~[#I@UJ}}t5h^L_p{D!h_Zl= VJ IP n~STg/y/{/"dØ.g5jJt^GPPFB0OC\q< ^Ql{U*tS.O zLe쥆A~K-o\ rJME׾+u,`_qZVtӛOi:᥷HCP׳]dv-nm,/8qI|L_^9Mnx {en X^Vy8^f- Μ}WҵV_ɳկEAc )|P^ߕ,IX 0{%S&s#p%Rj.^R\Ti^M̥&XT;ed|ijv',R!ݑƣD;q<?/Lʏ_~hߜj 0nk0q|6jMPVNR5GJEH>:!V12brAF ZQU"Ze'e jT@:&bP<Hr4 !0LHԺP7+fں,81I85;M<(t}d8)5‡EaL?6aqG1jô<%46$4hV!ƗZ[{]؉2g}\Wyl[zyr?E]X>oVd$hIcm;\an}Gp.,=}2I_. PYYy_==ifW͏miOa(u/Fp))sn9'VeGqH&&db O79վj/Hc[SŌ~nҩVip?!/pSK=_|*B 8zAUj])X #L?!Wwz+vPl9N6>n'O3۲}s&jij++ ;Fɭ[uYuͻw lσŤU MsAc|6nXtn?._=GHH%j M ="KQ8`ᨆ,sTa#'I z\ Xi2~4%k 0G -pԿM>&|wB1GV(:NJa]NlSV1F I ʣVC3&qE \PASf!I F:\ rBҁk@K 7%׮fViTۍN*~Joj>H]яCgR);0n iOp42=&kG_lfսѬQ\Xh2̮{oZ=P5Q=U5f]=3zNzf2]nb'5HP=|ҪG hbMM0\=q4dݳqKH|&u= >Tz 3m읓_ D?_; }\W F{\B\m$yZO@顳Aps^[_IêW$E?0MGBZ(wY=g!F͇T>j6: yɣ11;vt q.{n:t/5}:(̇g  <բvz9p8;B ~~Ň>jǣ~HҠi3 6 0 j]ib u\%#Ef0c^_;NixfQM5a|ho(5kV'ؓ49*jal1Hjʎx{XqsL Ą ݁L) 3I 'p$ aZ2uC bJM!惜?}?+:RL}&RdNg`KÎT^™2܈,ˢSEkHmj(^e%VOiMBM=o/"Z\aF34c: $p _bkkớOku͡ӕGyeq_04; ̏etd/փmX+v4X$@uU*95vjYVnF___>[%2|gWqv4vG894ŧ"T=3Gݙ'tA:ewGE2;D3Q5Oj\==ž=aQ39j3>||$b뱛'MF|o܏v=5<(EA${QPcve  `8e7ˊ=S=ހ*˟mԭpaSESrN㹧053O^阤eOy$[ SKE|PUZ*!F% [u.)Zwhvpl2<bsCD3yW 󆀵 UUkQq`2#+ڀ}wvm )%Fά-q?{Wȍ/l Afrw 2lƾ[$ywg߯(%$v; gcfQH|Xo 1 ŷ?򹜍tzÜg+T UA߿,[ )L UD^TKјG.;J(b*34h** o$)2=_(.:*2WAM+WN5I@y:P:8U?MڠUhc pu>kFr*Z|ӚZvrSk9T~ k) 5j;X`D{BNTˏm"YZKL6QL/?/?OF~0z*Bx- lIy?*e7g("GҸgT7,5R1^I? {{7yV{"ҏb b" F]zxhDl)7tiFyHGR0[ׁQ J,iSligj '.JvjԞ="=F6 bE_~$+, 8c%|OQ;)s^ T.a+sҝ{+a~,8ЛP'yirsuy *;22$Ůhgf’w@qyDUqDy>aj-*=tR4VZOBcl ɽS #Ke)-E~Sh垵EHL<G%V ᷂ 0!ե-8.K $VJ9$Y@QeբNȑGCOYc*{iVG(9klRKWfwAhJ ;Rf}ڑHj&/o~ǻxUD݌Owzړ+#tRg'UFYQO:.tL~WӊwTH6\C4$LiMHTI "S#-Z;=0ϋD%V:qp5pLp5&cK:.,bc3r m UXrRn!^&A^14Hݳvj$!C4HU㴜0q[3 =Qޒ!&:I>^dv}$%8{aMCYNMc6OǓ`?}i8y/';3I-}jD|?:v;Ч^o&@ 5G_7L?B;:8Ya&5O1 Q[{--uRkRVފH0-ruqw"\!" fLoN0+0R{䩑jYX9X)hKv6m= II6pYåxcpRxɤIr029a+<@XnCj D) WZ $[9RKVjNK)4J!.0Hm3uf*)c5 &z7j[%80*1F&8C _PYR jUH* S_ M-`4qx~y l0g8?m+z^|4-t.$>D>}|"`qm8~{(cD1S`/?{=ΖǗ߯!@Cx\e~7wOxtv:n[(: Nk%ٍ ~g8w ~r6韯NOqAv\JrE.̯ MMAObc8hI-H$~-!emR5Ӷ\aϩMNS3!λP#dk)>vX&'BZhɻ!iwlk!0ՐVZM:U`sc,3ٸ(m6l[+^Tk{R;X鸙~ߧ%ic3ތBeKʥ< :42wtdO7ggS>9,\D0+.p'dx[#EE gG:咫L#23Iofi$t(SNJRu廥j@5FIVĖ!]Ѧ˚G sܡ=oOS?Dg rB=oxH]u57亮eIũu] 1ڪ몣Q!n2iAȲW{ԳzV1RTU7SvןZ/-JeZ(SwݔP!rc&])!;㹫F-y }3#ƲKOҫY,X[.uM%8 Uf:L|:A'GoWDA|0:2c囪)B  9|I8s MHPe R8BNN==z%EG&*ٰ$ rZ4@F#oZx s9/ V\0J=@UL}#dm8?4&XLO;bl3;fdzz_.rLe憟M*ya(cw=(vzdEGmdz> }ESGBdv+9u%2㼳j9mao6 % %kNMȉ 9Q7!'&Ή[Bp+E)DbBK]Bh\LZ a9b%| U@^X ~"%DZB>r.hZM'k9cd83X~LqiQ(QGo=cOleeo* YՂăI_*@qgq{I \[Yy[PG t8V;㏁OUyv:^Gvxwr,P4y%ky< Fx0A;.:> "DDv1Ҥ1v]s'uFf($VY%&kY#Cc^YҹB+}! %qU 1|sd!DĤսb:}^rOW9b8BH86XRcȺL0ZFlz}1Y7]^_Ʈp:'n_CM B.A뭆K`F7R>gb^*tM xs{F^6}B:!I5eն"@L[a+쨭T_RR,.ﱣ*Vv!!;C:6&EHdUȴvQVV!SE3κ,Q [ͺiun5V2@cor::#Dgc˫xJ wϐ2u,/r2{~]_g9v6{pu`RWVnwgG3<ڼbNT!eHc{œr2:qOp J~ v) ;u ʼnD1Xϱh S%F;Yzw^ S9}-A3}.8GkFՌaO(ʼn//5rz.IXD{ZrPLR9}=.!Pq50 a. _L!cIIgb9 PuHen̳By(Ve&`5 #),X Sd,9ߣRHXBhVb!t ̋z%u^!ܻ\!X,./Wpyih'A !+e\ݛrEa4\>GHS\RD?pDΥ=|y\ _T0v2ARyImcb̧bK|Xdbq۠ie7;(jcgK[%HIN#KGCC\E}VvSՆWq$Cail>Nn=816f4OF؍hm0_ތGF3?7?ğM1~ ̑ c!WLO$|9z~XlT&bgD$2+W=+/C4ܡn0R닍)S*hEzb/)0*Ek8L( 26Ѿ)#%I)Q\ʆ aL5@hOd<뮌?=R%(_UUVy"\ 5͜s22`sg3E1Cڼ4J4JNSr'N;pV{y z(?ll=N?˭T02`N7DP.zS[1 % Uj́ls`7fY)9w7]J*"u9A^6Z(F?#:!af<U՛or`{5q2'<_R8tYde+ WI $Ne2nNDheh'KBNaJV#LKM`;mשQ k闉<8>"=Z -Vꘁ6::GaZȑ I)%k; eV1Wjyu6)Zk7WǓWSOy$|bRh:+.%ŠK>Hƺ?V%7s{;ܾ|kIQLr’)UPrkE%MXdx"PTҹd5j-C8e1s 9$-V5zKVP."aJb[a;XM6eE9ȵBR $ ĚC0Q(I1M 4aIhє`!Xo4կASCUvhwS(L^xTh!ca *UN+\Ka8+&ITD4e$M%JHjƜ@rdh1>BN:"'f#FhWcXc%$BH.aފLJQ"$ ,N4JeƔW)"qH)SyE؟s$vB `M**&annj69x vop92c#cN˲1 [:ErHylЌ8cL*4K@a~~wzAAy>l?D66-AqW "@.مiͅ*!SD# fL, 1X5 ŨYXLkT*0.GHN`Vxؚf1eqLatjOŸ́E JSTkms:&C0\&LEN2 h>F hP6WXvVY/ST]T!T*^Pî{ۥJH%@G4Jp""T邥~%H4F ItP$3(RWS 8P<BU%v=(Vf$H ""@0K Ɖ̄r"Ǐ\@GO9'8QU* ʐCfkRűћPA)x ʎgp "\ ]1=]hWl*eVjT@,8"('}e  1ڕܥ)*A /im zgŐA܊ήoJϮ|+CǝhKRuWîyU:*AGSZ '  !04f~+>4ˆL5̡7p:7pw8p#jdhTyQtUA\_YYeet Q`:DB~\0Tɣ|k8[=˿44752]p-#fQm2n9%4f?MǧaNgh &!ݷyEw9^DN4D{>Lw)+ ^>C~6#XYїqt;yq亚d 7`/:M?O3<-nX?L5;eu&W1>;a4Y^7(-?+Eu[7?|ۗv5+;W#cn!!{tC|@lX K+u8rs;WRJN˃O!W D\]>kG" Ww+l}oAoe{ (9Ai  Im*‹[Rȣr/mcZ!Pwk: m٢ x-i2aP@z@sŠ`Rb'ePhij3#@ $V<6 OpY, Ӊ%G9U8 -k!;XpǎBDnMD)En}`;]/4jsn?ˌ[Jno6q4qg6ͻ_k35qܲ)WhD']POc*XVWTh ^-njn T/hSW>v)o;ֶAl0J:Y<Ƌ?"w&I&T?鑯QED,$쭟RX\wl1G:=FY@RaY:nf`}vxԍN>2_] ]HC5JP#Ls|s5 7o(xvV6J`*+ˮ.?MSUAhncWLFgTa*Q2WdLoMp!C%nP&͜8t7^ \gD{]-u$-ԜǒH"mK4)fc/ľ^5.$ANi /> sAo&?V pm띢g{KB>/dS}~xfN ¨b&!|P]@$R*Ɂg\8M kqoDEOdX@W|\m2ϖrgR W Ք n0hhu_uFUto rRxEhcdt_}<W}*}뜫SQ? > 1戅q^n\P/sAْJۀ ޯbU٪o:9f|qo* SŊSxs<ͣomZSͳ_ޱ;CJ$RIڕym璈`wROYh4?NK;"#͞G `=YEH%$R҈dE$@, tbeJJ10Iy^Fˏ[4VA&6C|FI6=w:<So]G;`n7ŖͫNnw, ϥj҆{m:b}Ⲏ{O?c]1fȋ e@EDKqcY82fD2Jzf|:yAw  ^&l% ;sqq ୐S4^+Yz>'6liuc*.2GW":^ݺ9N;}C~i~fKφ^3z<b2YcC͙SgN9Ipq sG8NaP ?)~y)o0L mo5:?/o9HWR*RR!5iB7VF׵Ni[N?~/?)^WH($hELJA<!&do= NF2^j]J!q*Agcg;>mWL{+L[NW)WNd0J3Z?yϦE GH ^R /nB)Nѹ+D ӥ#+--DCοQõ_{_/Ϳ0g[f}u3XD+Ky_\?}GP$υ2T٧ +lʐ%f׿rTHgQh-B9$fFXP|3߀kNs#)a5 Z!gۻTFFGBХm:" eRw=:Nusj;гcŚd{TvJR"sGv6VA2Q?L_^ ,_j9W_[敛_̴ڶˉGP0 1ET*e3D2m4pbc1`kPig !xQHe z]U-!JߪfD %3ITȔ $*EebI!fc2C:DL5 &$\ Hub.5Mf2K+i ޳M\X_?3aϬgu˙PP؏k-2f/ةܾ,%ɓ5k[.{h3 [=ȖOEn&cjzA'#ݜԩ{9&} ]U<<Ƈ\u 38e"fո׭73yJwp˓rد(jWNנ#jkIéh53`ogeMXvƽ2`Wa:N>.Ze>A=KwWmC ){s ;fWyhP̎ŷ_h~ܨ4?_L]%qfn.mTR 1ߣ܆cȍ2}^‡E+bȧe7ZHRRA1!=UWٝm˅: -t93B d~uppP`k5}pKIo뭚L6p+Zh)r7̃cэ +<6L8.YM{/wAvk& ޘH4 Bo0CC C2y00B>l'FkVFdv3ress?:yPo1aZJ;,L@`{ =CoQrj\_uRrw^;f|`E WN(+GUZ.h$WydԿe`Hm*b}E>gvZ MҰP>+L2gcrwSLx6 epZ3}5xLvuߕRy(;\@ 4/uDЖ/;pE0`#\`V6v 36ga!Rջ M[%߶xnXdy%wE0lnF0!wQAzCB2wNG)ijZ:YYb_0v\Օq9q3aa4VbGEt{T.N='mf#K67R)Մ۟XiM|"xu-fƤ.%{a^hi|} I9١E 󍅿ww#/Pzb`^p{d`/'? 7;^7H]x˩5[oP ߻!e͂}2qD 1 "܂:gM˨Us{\;J{, 輡S_ ||GI(|^*K{7?~蝛H7[zΔIͰ! :NPB B$cQba]k}o01Dzai_3nSh;lN1T?ҔX?fd^yǔo0goVxJnY£΢i\uK삠KX'T FNf #'sɔxV1di(/_btꛉ/Z );JiUN>pF<\Eܻܻ+ WJ ck⊈} g g+g);4lu^AI Opvnq!q5?\[_0)I;yNx)|}c~Jp,`J%\`Dp-Y(-@kTkFkw?r#U8?+Rqb >҇T8<M Їw|TEM)—AuCsݑպHFDa^iy Y7Ǩ9[ l08Ok }zlFaX,Q[?U+@38|ŊjXt#l&Ը(l) {3 ٷlڬӴ&?}!op0L8nz{X4nܓt޳'8 f?EM$Mo= O0Z?7/~Bۿ { ꊣW/%G~; eoGEs"0.f,Qd3GdĚvſrœ=:z}ՓQ_zl7o;|{p=*?]4__0TcݻW#<2'^d1s.S^h]Wtd\J:eԈ/X Ge!SPβ6 Wk xj_a9x0K#pS:&cAN03j Q̽Z>Zqlm DN஢v vۘ:qty+-"fNTR+q<.78:Iza( *"QIu$W*-ʭ3vq3jQmq`}Q  G"LC!j$R|0霋ě(D"JR42S--Gr\)vz/%#WXX1yV42z"bFzna B(Ɣ *jQhB<_jIYr((h x(X@^9ibyIQTP0u AqVkG)%1z'9F9P1$D Qs"bZrdhF06\EIh8 ~2*-;"UYiQjCJ*kV8谕9MRwךJPo F빷ls t*$:B459p0G\퉴 ` fN*7zFAb)4ׇ8>Ji2{PU;?sj*Nsd}1G Y{fDЮ@V% dLx !ƢQ` 8A0XT5vN%0 &cL;y pFaG|8Q8 ,#ZGj"X`ڃeTt@+ݰ$" l1%˅g!tՁ-jp4 n0Nʮ8shg+/r]a'gŶ;3I(YI%7a~80u3JQDtaӆXͥIV׻qX8@^xZ7Lp]R'"snAb}gIyкoNܼK..ptەCN$K 7qAx;!:>xD!/`ܺe*X@֭ znR٩֝AlJ௢O Oza#Z))G'n1>{(z#ݯj?9(p3Q1nrMnc>q|%J "J1 QZKL7/M cr8Ym^!O0o@Ti| nͰl 钬kw.MdL_\~ms7~Scr|J ⩫[ZaX58|i_'~]ਔLF~p~̵W!VHZ&qiݼMf 0 kAӶy|a"!Q42^n{li68wƿaXy<.̼E6R9 8d4 E01MQ8 Gפpddt.ז9wɻuU0jQ9d"B ƕDd"1ek0l-- ;um?:@*kX]tZ{N\DRTQ3l 5v\ ø)&A}&jJ74IrJnLh UJuBj QIJPFhyIjt&P 2][#,v22|ilp'cAY'({] iWvkT3Dع 5ng4+r}0"Ŕ݄(dG\"ٱJJ{icUJKFeYi9+\]F҂#^"1S< MZ(W{R:ZR*8ى脤D DgrBԜiܮZ$s"%S$N..HCV/ooՌMi"Wm)0ϸ p{2xJAzO;B=y FvP(6,aQ{;M+<0|  I/k%$%Ak6W @@b1 Zc)_V@Z/<)LB_ ?5fb8$;w~:o:,?~g3{?o7Cw' a|Nr\RyO"M5[ZlbƬȺE!>XS|P}vQmu@Za^VSz?0,1^'q~}Leaއ~zQ#}*J|Y7}RQ!ͧOAm#/]S pAk uM4P[jĽ& 6US>c42Y<1&'/0zlknWΠXC,ci >Gϸ}^dL)FCu 0u:dĪ5Z==*>L(Ľ>/ =zQLzp-J|*[$c/I-PV3rAa<'{X;DK,ΆH/6v;>Al;r}}7iz[L Vv܅*#"qdX$ JcA%hʕKud8Z I[)q SEtZ} ^藜 ӹvi0ԕ# FY{7Բ ȃtZ-P8?f:U o&#PZ҅x{瓮 x͡ կSb>xٵ{oǕw6+o޾y\l3Eǂ?J ]D&c:^+G"Cv'y"NFH̅Ū6TKfHм)+Z jD]#Vsq v*|$rB`mn,OglɄ&j.&Y;G`X 77+mXJtm6tDw:t=Ii'0%AˠU"H ZA y?ղB;Ҡ+ bP3sOЅZ 2)&\/t!&KS?<<[mW{?+_6;[Lpom-[g0ƨ!g OziE2]w- 9W]-6ol͛Mo^>epK&o׭a2zQAƬ'Z-_Gtg^ 091ך_|0ΖTaֽw;(&KQ;ݾ;P夓JMV!DIᣉ 0x)z7eVQw}vi=٨.zr8yQs872%8T]#X5=|#hgcRb- GuWhp\sΙQ1ot*Ӯ#2w>D5!LIwr,k.ɳ'3\*B9J@M S8g5߻m ,Iѩo  <Lj?";] crq"= kUY*RU՗ 顚>C#n- Ƭ1%׀ IJy֭Slg)s)G)u#{Xiy$*v(.◹+L56>H?<6^I Nm߾Y;g` 1o@móBJ=K Ǹ+o]ߡw^f!%W{uF` C6FA63G|rh Í@BN6:_O5鈍Y000gT)k1o0M_uQ<(\(1^R(7.qZŹ\J :e yQTE[\ W!稜po2]8e~Ķ y!.A˫P*9jqn3z >8d\u0mxG,$Q{Rky2uNJTzΥTIFXErP&Ts=R;JWr'-q>ҏxX`t]1ĝ#J/6θ~Qvm6H;{OİGK/ {!+CT2P0cWN7L@6LcXŒDj88et!d 6RlVSiI0P `rW@$a!F@979*IY4jTtQRT *dUEO #{`Cb9 ҃ *?J¹;&[,R(vFA8U W.R,&*heZ|HL!I{T$w,SQNM'/He,DtIOE(JVWo~xS+Maz.H8\KE"06BtO`_>.Oσ_I}i >kX.jAl˷;],6q#.IV2g ꟿ7I`ld+ X>oƷun VX gq.Sf"쑩q7],>懓cY|u>\xwϟެ/Idq˟BLTxt+']+6=xL*UW=0YE$DŽջO3c9`)X%KfCxx7nQ Y-%MK~]u6iw-'+ SNVg8/*NSKl;[/"!ll-tnWfD&\b{s=V<[ل"d-@N3}rDu?'W0edYr`Oljy ~ߣFYZe{M5Dd`>-}GP#l 񸱭tܗQz3Ҳ`t]w9f@Uͅ[Lc|Hk" -aX/7Wy MWQ{*Պ=Ri=PܙwmI_p=TwgCקi`,DW[v%9Md({%)Q`$MkL󭆆VYH /+<W7J!d cQY 8tKᎉX];8;&nlM1y^up%%c-T)qT4`R{C1.O9!@Pqn; )'s7fp|ֲ0yƠ!Wl)rCDH? ޝ1B2)uܰ2&\4շDcxߵͲ\!f-goqN1y,$I}!nQHo05}y^|$yɍ C F>[H{]₅=XBc8!P+!:w֓?Gy(cT(j7m//k.=LPCCj 2pٷCA$|Cwn1ZɻcU9W3FO3?O: b, }}P& +͉3^:Y ;|U4.u4d.P*Q*mwW쯘EB[\x)L% YXXIJ⳼ kXYT2֬RnJ?jGKz\EMM{6wzs!9}ݢa8K,#Pm3nپhgWuQc,kLj7v?WJAe⢜;]\š'k2ܽ߿^!4\~n^\ϓb0ӊro3u~#f7Xs5 ǼI1](ׇrQa{VKhLUz0n!hT bD'M B{Fj6$+$04ji4%%[{wunxf߀6%L:mÞ)E۷_w)ovN w7vj?sJtE:OFl$YSr-򬤥UfXa2T:G'v{3 ()#+A@eyi:6ѫneɇex豖PÒ|cdV~}g% #-%gh,9| ,1e ,po=aOL0>Aw 7N 5y&<5*Tʜ㬤Am jebOg*sLKa JWc$t~igo ;T {P#nIvc,&e"C=d5lz`}O*#֎'*RqT ',<=bz%f & @)yQ>0TʒWsb a1g₎ij4ZFViC˒hnMf[ 2s$6 Otn ^!;*Px2*jUnSz+o=j+s]B2Z@!ԧD=X-ghޑD2]mK^tVΜy/? D*&Bt5o{s GiG4*C2xzuId+MLf"#>? ȷUm٨"%dB1<2Ðf$0 ђ&+mcl+؝<kĔSVzy1nTA9r?fzW/K.}>Wc;݌"#tO>99#|9z _IOZ%tW vv|<9bؠMOIuK^I'3G.,U+B1= o-R>gi#mp d=:ܧMrZkwwRIy鼱VN~q@GgNF#|8o[e'S>5c3*]q!b ӡ6_2U 8~zҫ\3#*=6(_Fv' q=Xo?ad 2ZJr@s1_L\ymEq݋[]ztyХAWuh]eFEJ/d>'9-qb)jY.c[Y6O~ ylb&YNj9J}Ћz' YO#Y|,Oɇ&< Z8x@#ϭe9QA=ZE$W@,ꅖ r-E?;jJ0jQXTA! $ϵc1De5LC^*`s a둊$Q񍎤M{5C߭›I5jKaBkFG<E@pZp;J// !,~g.x#Ĥ.8l!˿ޝ"ngM/o4qZ'ryBM PXe X b+T(h5X>)8(rnoIl56d16u dJ 'kV2ng\rK{tRyDR CsK*J:II]]'Л-iqAfy*#{l4RD5!]f9ge6Bك!C1&7?je$=yu3{1Ϯi^;VnqO3c[U#̈I;>#u3L#SpdvE;ky2JdRphi <f{,x33:{BQxV`T:7K#D (o)O`9߄gKTm iV梐ĖԡSLrfԔs" Fcr;waQRV@~[EvyPdAWYx!q1${py⥰Z*#bVPSo.(L}PRAa~$匪!| ԴXlvJ#fQ"4*T%i+_Լ%T` mKmZp&B _ύT;Q8JZ\C'(qut8Zk^fý:lTl'\MqK)s?i\BR Imuk ocoMCY2>OF?禒Z<@MҔeAKٖ%CuO˸]ލϓū|I ijדN՘?xM WPG:-*L Ǔ\#shoJ;{}Ǵީʏ7rzҭV5⑪qwNӱ͇tsM *g;Zg{qWQ|p{$ev]+Eʨr5qiPq׳_5 <uE3+p|g[[ʔ̖b&7&~˙%,Nc9lR{K~7nE.\_{/h@pxO۪6~+e$r,8I˵xx)2d[r'|FB>\sZFGAW|`,disHiD F@,9$w )B*j"K KxD3\)6hz; e84lO/qQP+鱝#?2t)]r26tn@dSpk=:g@!M悻Q&gF &)w6S9iK"c+/.>B1IH_4Jڎ  wEA( d62N9ma7`)XpFκ`r!g1 e a 4LƝPPBN%ڞCPEP0p Bʜ6 a7^**=H,\`4ue" w޴0|ehʺGuβqmܢzE| Fhlm l'T Xse$#5?*57 Xsۼi]Z ޠXBkP SL*I Q:ָ?s@{bl J(pLՉRhƲx oM nT$"ރ^KɁ?nx>K341.՚2sdM܈ "e ̪Zr2ϒ޳d3`qĬfۦ2HH]#o[9 V09y8U*{9gzC]<:?>Od{ywm3$hque5w(89t"H %JCQ -WѕtoavM_#H%ݦ&bBm[M䋏v%(·~CaJ6n廅-P1O7  Aӽl؛"MFlrr ¯P#z]h>|B*&CVޭ{ҘS?Ք H/nU5o_G1-taOXϏ?TɁ hF7zC(a+Hz*5uAZcBUF{;mH(-Q&˽uwLRk5=y.bi׌|J#4لs+qR{QغƑ#"K3&,eA|૭]ƛ SּX鞞,/`xGbȪz dPįt8ШW9Aʉ_sxW]4d:Tp_bǔp`W8Xy!i "5ANсCU{?T.: c, b,1w"J*6OAR0YDD3: B5h3wy8Z9cVȺN$C°6zE&EMf-$%JrM]op ѡ:u2(DRv֥(%no1Ҷ/htb.4 EZcm)Ц] Ukfy?O@Pv;MYwC$iP" (MfV5ԧnVF,T %Nr4he5:}hJ3 ͻmduV*(Hz,hvsǻh;?&q2]"_ɚˏpsz,G#_9^#Lln]}-ȺGjrk#芌AHLFM'qȉSӁ9塛9{Hע4Μ@waXަ~\ Gotc?(s4y0znƾDNt!ӝ|oDfS0JR#n.uI3`&:"%Luʫ3C:*ԁ+1+B ,:CӋ^#ɜFR? CFx8רFL򰺽TEL63 ʹR3!8šv_Νh ![k&u,aspjpS{/;l斜w";zn%_O5r;'ʓmz=)H3@D]k`2|r2|={L/NȜ\O$;-N.Fۂ_o)݁K_㖃N9TEel`@X#z5Px6 گ{?ɠC#Wy3VD!e e$@N!%NZbA[ (c14e?b2TI$E'Y)d@#an뭫̮/|"FdQ2-+/r6:fח9*\"B5('_-&С AVS@$+) BA[R)<[hl Sn䢞3X8xm Pok69*w⋰EihR0 #UҖ;d:?F㋑Jed޻\. pmv5W, I,DjآZ˘gnE ?9cFcKK>Y7GspNkeےh_ӡ'vHa9=_dmα\ ȡL|;ԩGϝ=|)ŧ8S[4RĴ w7,rp @宭lyu{9Bc kXU7_>/}<,i_/o} ?J>\ V>^.n~]^ 7:\;Xczc1 Zioe(Rl 'rR!vB>| mm.؇W 늟B'-KhtH|3ۓKq; m(!Ywbc&:[$OW}0v{3`&L!ONI=ir@(hY]Y^,B知~' $zNdage3XCQ<jgCTK%wOwڃ\ GG=Yzϣkz7ue72YK⻫z͑YYμ>ߝ\Րp߮xmW_.jB8?]Vޭ& }*.?hvq2CikxeBRx56L_?+)_ǧ̋(W4;WA:o ws#z;ysyܽMQhscc!!hbfɗf|f[ľ_'1gUZ7ӊH~sh.AB>H$+#KPc79OVA>e$Si~wH=N6[6ԣaT}~Xw*(m/?5Ѣ -^R/De)`&)}HozBWuZj-3YYs*! UWNz+"icGsR&&ަݕX>Ln=nSu׃x9paGxJ.աUPC:|l PnIQGgc.@0>3OʪF EL4'Y3QkI(F"9(R/|S8ը1ӎfgg*6`. M v(DeUF'u'z`ذo0vd.*Cl[[V+{-wQ ֱ+鬑`0!%$ןGs'E*C V%6(mMB=ϊ"7q 0 ;Pap= w؈{θP hSEMl:O%{@1(D ~ۛ@5}O|B/ Hαѧ)ZC)D"ʜ!B&kT+b_^{\DNeb:*HW]BL'< P ݾMkC$RhMc(Ly=[5k-v%2T64& 9y[+$Yٳ7g :#o㍏S]OdmTHD2 y*X^:\Q:gKҞT#X^T;>.P.ť>e|AeDu8L V"GFf3H<=V"u:ĻQH37 XSr$ㅐ&;qQP7#3hK";V}άlEDMASb'HǜُD}F,' Fq$ bB"#; =Qe_8'(+L*9G#"G-*G9o}D~+e[zX r/Y{ox73; bzvO/Qkjg qA30&.g?[&K=.W9s7yj f~NGDYX>.Wa3lS/37WWYxD,%ᓻ>D){ړBLRL#eftT<َw'Nnu[sCh9LgKhޥUWzyoL^O7lxKVY}^[gY,B{ o~k+{xE&]}Y\Y;gv+?Z,+T,ع[gɽY!Q׮t/7Y%Zܗz[ԼUۻAL񁽫?.whixzw~π-73[Y-< pg43|3Rl o}ʷ_.&_ROUAQU6{K;8V6.BtC; щIo&.q5(aBnԘY4x* +҃E\bK[RI Sh"b@#aFC?k9#cJ7h 8C. Cp@y졾|&o{~!jC*m}I4o x$e< Ũs P:!ݖ;=[ZaFl2虭]͊myH\bt -[|vȀr9jމCԍyM8!~Տ%{>nnW5Rdkvr/ ,< |ql7LĹ`-iU;/Ek7#6ێO\n3y&w \4/ΤX1L^e '5>P~;~L$4ԗ]@3>+/CHr^l4tuCg:y;]x`ߦWO`'[nHWzc:]-i'w.ѕr@.T֩e&L5$J? +@WS+#gg%heq N?-6pIW./o%KY-Ny4#:x4pzuE?lr#e?ZlenhtÏ9NO's6*wn(zA.ogѦ`a.Y8p~}`2~WQJ9"K/(#4AgL:*;ɖPٴ#s0)[g#G&L%hM0+ / Y+lPuR n>[/R*)%U!۠>y5EJ"I)b@Dh)JCTAĎ #5DN/;Ag8 $PKΫDV'J볏*J>fN=e%E12 ~s5eҋ "j1X%SZ`z+ x2\A)p) VA!6lAi hnq>Ka/CENɌTe.,\eK_$-&{]Թ٩?K ZfܮBES`cRsQB@ H,SMC~Wzب_/j w)glޥ_+ fޜ(ii΅\{]4)<חg/#5 Z&tr cgrǼ*J۱}z=LF!*IqQ= yFS7`eҏ1Zu˞dbG,{ـy,v47_̟'cL?zl$h/|Rv#mr:4 kBZ>o]0CRC,FrEtqAK-Gz`Ǐ\үERwݢAսc@@< (a'h_ <f2ۃRY岀/RTU6*DM^yzw kȇ%]wJeTկZàɚ5C`87YdX5jS߷EFHr*>~VrZ+PMlH$/ĊN*qY-6fB#Ej^w?b$mt £q|[mϷ}\O/n&YȌI mc2SdI&!qqhlNC^y-2kn/h5M0qoWYͷO[ş/-nlR{ktVMw*{>X1@Q@rr,::٫$u Δ1PdsR8SȵvGqA]%ג*!ͦ(yA@dm|q/U;‚1r1FKP%_Jmɭw1F /Y Jk0G"KaXwݢhU]?߱R`t`ڌBS݆I`d&VcJ8LzTWO]չL=$W؏ x"VwA[>jTW-킶IAR /u 38 a[8q; È+-ISMA*MEn#Ȗ? XX2Mv \+md:r!}Z;$E&}s kG;RytrCC}~ A+}L l>Ԋ˗q<\J+O>.lsT/2v?/9]](q<κ! Z$8_t0sNѐWg.E-]=^O$Hwg")&9"~39b4MERM@{-(ݕ Oжjn.EZ ޏZ$QQZЏ$_ˆ$#%3:ӪQ.f-8dɚ6L[T>ye #Fcc՞qN>| MW9cNĢ SAbxxMDT6{ӥu'pZ;L}fۑYjeY2C;ֳh;@fADkVhn|(9iDޑ;+ \f󈐲َ| ,ϕx6OCMFyL<ȄRa' })nyߊ5[3|)J4aT"]~N/Oo |n Jihh@'V5JDV Bd*^hmYcI-3EljС[f u{[Ul">u U!dlGf&[3Ud4edGWN<6ZeeҞXvZ"}_i.yWv(v {wO gȚ 0_h̷ub ,`^n!9K#&P ܞIfwb9)QtRRZA_=h.&xNi) B|s8x'1jӶ X1EmLIDPvSV& U%!jUu,Z1 CIQ0812i"DKuHǀ+<Ns3eUSz}/?ȻA`  G U B0W F$_jіMj88`,e 6oSva^OPTDV\jC&$/--XaS(|rU;ћ7 MKן4--7MK>\ڴtMFeědqzg&1^ .<<%t[XbrzjR 6.q_+Asv?m2S+4:ۻZh8%߳j.Uўy0NS"CiD< bJ|r"h)ؾ.fNs! @uGwMj\3y\_] pS?\]@. 1ٻO7y+~L>eľ$}*@ahv|t3-"ɄʸiJ ](9 Dt6*a٩-|;GzxۤSpc SM%୓hBZ eT.wtp%?2gw]e_~[m*=~v8yJ{C >vЇZn(vЗ*8v#DC؆M5 BCWžWˬ15Ah&A3\M*{3xwח/ʢpj.Ăl$u`? ƫCOe*xgn!F~8=:2Iu=bleJg;0zr/Φcpu.?Y?ԗ2vV!1(e\TKEPM[eJ^0ZWl9>x(=e^b`T;^聃8@adY ŨB `1^0TzAc1H`"Xpa"C76iϤ)"G+ 1-lA1S(F3\V!mՍIǢqY+9 Ԩ{&biMk'`"J&ٳԤKsKxa.Nf\hRvS!wfEba:Urt垟))&$eJ\dN+’3Xq0{wc7F_n<4ky|/qya}g3ncqf &Hq}!9A@d&paUIÜS F %#UucBlFd-x^Hf9& ahֹ^(;z]:cPR:JK`Zx)3KVHL!߂1f: c-DJ + f`oCJ*Ӿ:\v؅Bmjp-F} 0ϧJ&84h,UA;.OM|z=0wq0aY.6dv@"s-(ЋOb7rS<h}sSśWURO7^-ɩڋ Q j\M+39@ 7`UBn"iLSk4'yX-2a;^>;ޅyTCi)a !4$q<1y` Θd~W~Q8IYFH]eOwW6%/\Vlۋw 2n&d]/ʝߙ6ln~xzpxfoԅRXsp)OvR/ ?Y;e! fff9'rJdp/_v~" E+MG:E1JW?Ħ]wi]}P?ݱZsmt9]вHӍrXzrU~J 81#ebQ&`z$o>,M~&fP Z 'u6Zڥdh Qku֩>hpڠ7ՒJ\ՃS vjz u.; IkWvҊ[T6|5D*QvjF\+h)j;2'Н^ {DafU=U2J"XB)qaYA 5EE'ԔBցjAyH2cXe4 9K 2 @SL󤄳YO T(Oz+k#C3:2[vD@$s 1R=1C)z?7#w#}tERH6DsōA !Az'NwL8jQQ͕aZ-\Eo'r𗁒$„B)x}"_(]9*61_H]~[CH7.gWm_^gk77hjϗ!Uc(} mªO?bRQ|;Y/} ߨ^Q)B"`ycEpZb"c>1Ӧ'dc Y-B+uI HM(#؝oO@ңm'2Ce;c [配 Hu-d ^G?" ^=w!/Kth,Ʉ[$Ra Ogi)({-\)b]L6C{ff_J}4OF( jH@QJ%BRB".r,WdUן M`чK+>dT(vs!%ߏB跍ſ7˵M\]l81FBWdR LBܬTQ&4|Zt X^QZOe!U(X-M6VhtFIDL V Sy=#8_^{ iqbC57g3YEâXm@$E5C>:%+SSȼYUS?w;TV .Ez5qCIh)@Rw2VB?JhV3t[qp(6 -\E_Q)QZ7> '&Af.NO ^X"fjt}`_c1 gq% RӔd*z } IH&Q ͎cM\p nshG ȧ[ӈN]u#'#F03?h?\O2e7+ w}t0& k^&pni5I2 /o]`Vc?+ycP-[17dWI]#[`bJ~ioEIjOUr_X{/[3D0٨0^,b۱t7zqʓ_~; {?R&߾Lm1Ov??\(G!`,i8aі&1TG< 5{[[YB#L@WM_I.tHOa[tC-Oc!z6f(`;أBKARYEңB [-@z6 #7)9{7S (bl/S}]|+bV ڍ=3Nf͒(E*ͯ6iZp#I#w2σ6Zz=Ta]O(i,ccGR`0L vqPT)ƭ˝ʻҞrRfڽ)g5w (휧CW{GxQ: 3G1p+?B$J@enèt_AT<dڐQD[%'yTBG.0`l &9⇨~C*=?}vp;h"z?bФ'dc"Ӎr;/!QhRž~ׄfp23_B敨<ujCiZL#R) "fd?xYy '(R\7Jab` kav, d^Ro|d fC`(R3_B T2Τ(&ޕ6r$">EgdF^dk11O !O[Zݭc{}#IUH)$%b*B\-|ֱ Ws}J5K_HqI'u&ߙ9F4b,'HtYҐky [ AlDNE)`^q"RrʲWU|Tiӭ=kX ϟo> S`%vyoJLx_?56џ~b˛ۻxmwJrYX`"9+w{5ȳ2U w܏ 0P;^]#kܾEaD^]>wMa [+Q0VV#еc$co(cՉ9JLąZ5Kh/~: ^ o)^0v9@ZڌWq砏ao  f>aeuޡ.,X;T]\"c$. c1jF]I=sܸ9DЂ8(1bnX$avP0:ƬzGTV\z6`vR CdښU.WW͇*ҍ]rKځL1C裶! +\:^ u*X\L]Vk`I1G{ЍvH&7-ڣ#tWYfprȩ7-%13HAK&|#)kcjOaAa2B ti ҧ5$lQb勹`TirG=!_E>^=4D e"J<K}IU?S#8-W*BBM XqP&^ۚ8xfT IUR.% zz2iGKܛJk^MD'  I!X}QYDrwTWFRoR"$D!X9b6(r$haD1s,72IZ5-IC5)QyW# '߬AaFQ)%? QG#ҧmݓB6'nj'( ZCu#JQl' !@ .FĻHPgHjy$Q}MҞz_q/Ju 9o!BvROa\`$T+49-\VP-1 *붶D "!Z+Û*(V!{H%h',h-NT=ӘK)o֥ ]ɠatC y[.T.sPT?Q=-$dzN+m; oT:F˘{&3U6ȠreD5z<#FezrØ3Qf 1v1E/+S1XX 9rNb59t J)K){H:>18FW,D)RYOR0.Qjc%Gd#9X_%g.li:8! " :Ӫ"3L FAZNG:(!!STAHZqj`#*|j>E:Qg L1숫x_&lK+9732݈U\m3we%IcH:SܠI[}q1 [; t%zo#;>ߦZBKe':+t**'io{(:o!_3nnDUKqA0zs:1&%oV}Uկ=E/`1ך*v⊄7RLJo~L U 7BVoլMeeZ`:CP{u)&&3%hۜvߨcn@47& U'=i!1~s֙ i4hРnlRqԬĨf36~]n|)׾0u*mhNTFE\ӪI˚0 8a [W)Q'`z@L8FXPrvS2;$MJa%TK+smJQp>Մ !(V\U[1q)jʪ^^ڂ}뙒2t`@2'i56]Ӧ[Z<+GŬ:|(gh2-)Mnv9iiZ'4&l#]BvI'zfaқˍ95V!3t_/_q= jXSv4`G+a5nkVL`QS[6чף*)Fzr*>(lx-lϛUUe;.AY5;)BNZkvI_N*+=,_V!9ϷV NGJ G P|o\.%}>4ސ4@1ŚOֶי0!~g.0AGJu#|AkMC|j S헁ֆ[ADk=-˼͇wÉUL#LsuS3\fݮe(Ul I&yhWYuNBp֟1M5KYR]II1 VA0&mR@jfdAdOUnS5|V0ƥ8*3QNY1YA tIY-KSrAAJaATdgƍDHPʮFϣN5w04IK- &uÙ &ذ&[ @9rzrqīu#J+WlUmeHZ*>8/(77!9LFo J΂N _0nC EȘpR_lU/}uV@Moϐ! oJAaZ>-yvKj=tYs?|uVaBHw)ӎpbqE.JuQ⬋vնR.>PzAD3 yvh4jKjA1ݝ-GF 91Egov jƆBkQe\EM̾ >ELiWg{R)-Ja"w$I̵w# Ҁi^U}i`-[يJ@'kzȇz g!֥3g?˗NjD 䚎Q\Mgxkq"&o/UkẈj"\u>?$gkN 3swO{նНu&x{ ՓVWϮd UκU:ZwPsJIG5(N8 {كpmi41 ߱=,gǙޓ?'W]\{N^19^ 4f],D¡3׵ݎZPrS~s79ix;Օ腲Ҏwy^͕=?'9Gb)}NVm9Ý*?#ݔlt1?;6`L!J>B~5nd- ~Hd8?_aJi-aJk꧴cqRͨ+c)܊aP5jz;JR8ԖC_yL1;чm}ua,=B왭qކ "[K{]J$t z)4],Ib>\NYf֩#^ήih6ބ-zV̰?'Vl649& ŚsL9'B^5}xR%['m3.IG`{}zZ< ϩy*U+:h&Q?^ j.\'ZygD< 3'EM(VCVJ+'?mAJYם-VgK):fॢPTq^(yQmm栂G7qЀIpN.lejeP"SsPC4y㿷Ce};y_ zJ ފ\H1>Q^SQѡ/K13{NRKo"Fjd/e1K䕯ßw7AX,em}Y_m}KZ~paK?K6BT;}sx\{a\5L4- F~w)nn֊<ծ.J}]mo9+B,!ŷ!Yplfvv[ȲeqHI[rK"M; n"72DvnrU-ېHZQ!;qwy7omu= ܶ=92E$kލ~QoW;.zzD <8)C;THsCؖg((ʁ1NyCz"?V~O(|'Grfwޓ5I (h*[?Ά$!Q!hqy!9ؼtmW'MVꐅ´̀7SK(g%ur)~ɫ}$Cֻw:3hWzh[}x?__ l'f-w$ú5:pvqr$+'{Cf:rt;x-$Hlp!< r3Y䥐%Dvep\vEWR#Cs'TSb h)Y]JP-}^ٙЖ8#wSw@/IⅩWu#4A%5P'=0[h'}aWv>_,4"ge\z@DQA"Xem`8O3Cg]qWg4h5YJ%| Oî[`歉XAɂ$zlr`j8Me[,'L=ʄ&*~nZHQx 33B\35`jIqGz EՐ0֓cSw+GmeY3%2\ӭQ9R^/ ͮ79KqU.{`唷=F caȴΖ1`"'pPZrIgmA QHLg3%Q\O=S*kxYI2/0p 0n)U)yCʊ3 m@ІYo n8"Bh0MH ZWJ5Bz^j 4#ƕOۡe;ÞgeOHʥ|JE]6 DCKJ3"tSU5Sr7,+ i_A eEGVVESJR̎"~VS?' -6(tYGK <җd\1ƸI^B+"smO͐H$,+v5jD;>zZMd3 51kfng{9A*6_I!(w6CvݣvCeQH#o{U6ԨSO=r+RHF:aEv*hns97?G` zؗc z ODŌ`J N H:YCjLĉ ܥtczXRN\p]JelR 0=g5rZ"9='ܛn6W 5Pp`2zvr4CF@$Kմ24bo,;$]9rHNކ29c=:wP29]' ;{O Y~my뎒^/7D=CssbeB y0)Tb<̇R~uY>\.C/ CdS>\fP쁜Wd,!AiɨU1-"~e<7| 2XSӒNQpbF~ϗnE |i CZ`:n/FMY6-yq7Ut9͗hGa셹6D=75~|Ex;(y}=gou8Coy(|QU``J}c8sp\R5 q֡oU횪}?ԣ g 1wcM OQݩ`.`ĩy3 "1윰|L- {A1cYb(ƹR<&yY7fX J`<sΎ7C)#d+7c.Za0#AÊ7q^&x D J dLeUBC9mSsP`مyJY$u~*_vڃ}ƵOIUѢ\Uv5)PZu% DJb$)uYgkO ,jbkPji*IH$e!5eenARVpRb X''.V=bBs/"W\ x*:R1A xZ)۵`J"o}\P vI*0+v6*lm/j?3s3ʵrB?viV΅S|ضa2 [E^}~JHd-4 TJVf崺#;c͌EMTt-d*4DiZz?ZΨ :8!^&Pn:HD}c[cޭz/(SNN=gӗYhqePERDqG(6)" FdCRi5N}49a hd0{HEa$Z4l#TUU Rw]axtk}jDžY&0ʽ ?-Y,dj2/p#ܴP-.ohX^K*9!5Rִѥ$BTT4eQԊ+J_zaé[l'}_R AW.T'f6;p\slZ/ryrY~Ø(*Ο*;"ףC>}TwVJ#RԬ "k VF*1PJ54`sջ 4A45hSVB#*ךUIie#&QfȩD+,I!5rFN5Q\€pkT V ;* h6n$ɆJ7K-(e!>h>OX;W#E. c[ڠfu]QQVXZ33aLCJhM}}}~ a}vERq‘uG[FWKH)LzhD7@Hi"ޚ$*AA鐈#k=v5]?39xp&PxJmY9ʿdnelj*75[t>Si`9Mm!}T$eT-1f?]Οʻ[e> Q"6'Z I][wzX$sOg/h&ǮCjgsg'mE7>Cm։xm'svo_TuߚP,?oOO:9 rhIGbڤiAp[w|5yɋک:i(q>\݇ttyۻ>ker6g96Z9Jnכh:L+!nOaYV[S|-87F;z!wݥ~BX`,Eyr3G,@g (Gc*Ou}hz[ /<9jFtc0GݣbIU_9nDo769~ѐMNr*J3DE BԽPrJ``:'4{#= ,]F^CHp woF#0w |6*%@,Ag@ț~:9bc+cAFhcYXRe(0|wPTr+M_I{d /9fTё4&>h 4U;C-Gj1uOdmml{If;?5ڋ$G0 |X>=W'Ɖf哉Ƣ Ԕҽ5uΜ(!/p2+-U]b1hM02BZ<w]%W: zΑحsHB3Ξ4EJRXզkim}g̪ ?p[3vlp۪⤽6ȷ;Gc姹:v!iݛ 2Ѻ 4^@'Uk[mg6cO[ym *"Wi 3="5i͕v~<;++1U>Iܔ3>wn{r{UXmʬgn!& CvE"Յ-nk<֖ǒa^=VoHKT,p;;3ш2]sqӴV/z[h{O+z^< bF;lBItY\6\ol?[?w]~䥮TڮԮjJ*.je {>kr[Vwȱ-C{@@^Xyeyzd]ǽ`/*^q^z1|ƃwV$)rɏ\Tu"סTM3pԱ~'34Xk41aq9bYqS7IQ:`.Bi,fa3-2bZƞpB~~F,@=F>p)Á݋hxL}To9s 8*}F}s[&ę9sV(Z239 ԜT)DT+fԤ\8MHnV2,"М uSOsDiy ]7ŒiW?m< 4s~務vI~f_6(DBWEAhekWu~O܊7Mx#:۲ΧӴ aPnHaTB9u5fB ?_hgtPe`$q 4likv߹,::9iOK:qA5=`tF Q}sxͣ 6a?!ڥ)0a+IX7|l&%bJ'JXKCb)e#"FibFbEFX%=RQ"d,^]rUƚIj#B$iXH%VIJeb``qb# `DÓfSNc*5 (FR$6ab4"bbLZ% N-5񔚌V0Ǚh,I-N2uq $ 7!G`A<1*0a!b e26~0}Nsj@ȗ9m7'VJ#mwEkX!_R=9]#*kCYhָ~hu/ M.F# +! U+xN'K>0h6"+-8GUPĬ=GWB r?(.1TIyE0qgWA` ,<#&p ME$C+ = Á_n8hƫzdCԮ9\f}8JL#K5YN:^WmDcg ngP,WYF)iu}8ժ5w:HʩlӒbT.nM}z/JvEWHQ^wNW1T3՝_DT&3j?UZw5IV'F: !,v KbHJsx'4 \s6TTSN}ܻ;䨾nU4azoε5-12 }.jΥ )]D\KPC$.;]>xNs#.\"ӑ9'^ҥʄӈCz2=^lŠ%8z谺Ly`e/8 LU%B} *RJdicitVR 'JM>3Ɣ:ԗ䔎XۃКrW?[BƊ#,:U?+-?gЧ빔do빳xZS{18{׺5^)zH2_^CmQ^xZ>WĤhܰTy0m\9z"]6a!*I˓ ^el&doy.{ZjrS<-B]\mhDy `hKt+FwaI[^BN9ΗDQ$^Jl Mш2KcMk*KWc3~L,RWU _} XW3< l?bu]@#Xbjk|L׃Ń 2yƆrne:[7ÿ!MgWWZIL8I~^q8lyEֽf0=]rXeSD䏛:Jx]CVo^@6W:}aʞT"0M $ Q(wKa޹Y,'D̰ hNƂqcEYAJowdeR;ca>d5.?;_aO@6PGZﲲμRmDۨX>kOKq`")I`c B,pa9JdLP0Dʄ *_Ǣ)9~a-6R8( '\Rk'Z8!Fkq!P:QJ(9y Qz:GGZ! c}L)_,@H Nh6b#ˆƇ@2!'`;#׺DI#stH]~K +E7dwS9eV-2 jӈH>&M\ C>E`^HAe<18HRZ6X2|OFaዴ DEAAH( p|aicBmZDw@ŰL\ Vz "iZs?vY8EðQ9-o2Qaꪋêg%$.aCt&bJӺY 9!om?KVV{cc|%D-?)1uN +mG]}D93{ rqx>4QbPPJGTPXE0J f c"X<zp7O3d͠ـ!B;= /iA`J:N- p SAXIp~i$1HFJ^!N 7%Xm±Qc QeCs2 966^;_*]C a2vt虸x$|:\"%b$ *Lb6"2 uDB+A.c͉ !{ꐏXؗa_VIR!O<.FR)70#1X "ѰDXaH8%0Z,E1: %a*!.N[dD9}#?,3 Ղ' lHHx)oԎ5AsшbWg@86;ЄIOAm.l#SロRѨj{g湵Rgdcf`Q ]q^V SV\2)O-kyu@WA˳RWYQ]1[ǬpZ7WE\saH_'bVRpbP=rV>]83Eֵ+i9im壛k@8#OM3>q/Ӭژ|L| lxu2g0z_z9RQo!KK֚;̺PH6ߩ;H󷾢`[ f?1ӈUDKH-jSnubU)ryToMyVBTR%Q.Jqkwdjُ \t*΢v͸Y9, POՏ%sY'?e}lB^ֲR3¾=;ލ`t>w;c, en=[MM^J4wө*ޭCLw-nXnMX+7-b6FI.v^ gN:Y!ڝ\zbq,LnW_d]rzU{ob=BYv#) ޔۛm(,[jh'&Ҩ3&{WUߪ4}ba?+T֫;ev6Bݺ;ypbLpk"\P R\@!GB3+ ^wx{E%c*ha}E)~(P)Pϕ;^Zu+e?=l.eϯb4T^Q!T?Ǻst>lG緐^YóǺyy=)9[?{T)٠Dtst.b[dLU{馢rx} U6:~%~˴RS J5ЧNE]y E K_Fxh}lqeN50]c:@R,s|Zǒ۟bSR Iy Xu ѥx>=6b%.@)5M=<9e{`IX*V#:e;^8)2=4'H1͍eJo 7ζtbTcu&SI?L)bf[™1`!6\N\kՆݤMG6a TVؽvfMvS۴ q[ڒ\;N>Ȩ8fI`TdfqX2ZB{#x)Z.հ @9GwlP%<~E(bCq;2⟱B3^Lti^["QwVPx![ǍqI0k2e%.`1Zu $o H˦uh {<_qyqي^̖TU4 %iI4 sٴgpgw7BGXJuRw]!%I݀˾~4*X}[M! CE!11 ! PZ)kLFF*e @S|{*53!L塐A1E V +JdD"HG"AuTJE kT޻ԓf4D'v&(w~9`k?{Wȑ ]003&$cY$%eUVc+Ȉ8ܘG/g=;; 0}͊i(ʩCȔJ|cBk6΀$i (dJh/'?3OVo4p\0(cILI|pw-_#R9Vu:oMN|o/JZF{җ$p&D: G6P# sg'Hjj U-m4ݣ6\ǚ6@N-+? ge+ DU}Etor`)%+Y[G8̖ 1A8p"h&*P5 @ R#dllJ:X\J,ry`gW(It=ִ]Kv>+3XC($XqJCj%縿hVU_u]!rXMš-J;]-(6a-Z!|ZNts Iݬ E&";Qy94k)۽ DΣv'sߜYLo9o8L$| 1I;/-5RˀK8$f8N"R3F6A%(8e͂S\*[J\KC4qˆ#I*lht4/C^P 8X,1NqO| ,Ya :H8sINF3G:vl8Vj%p;oVD=COyϫaI^mO(#E~砛Fpg,-`uo"l+\h%ݸZڻb)_|c\w յ_&T=z$`d w,*1qR‡2H\t1U2h192BeRLp8\i(Vːaz(X6Ua}(bM2N(‰F( 61f,:H&G o<'I4(= ৻{Y$ #*1>)U9 {]^+sU/kǾGt_'k{囥A4J[o`gÆ?U|Mj#@XRʂ1RP>A3ͧYmtze2.W/c慠LMvNR.%kCJ3u=u jAt'kXdA`PePǶʄYJΜ&?DkyJS젞1l4%B`%9潅N<2 f-jsz _ yT3\'w-pGě'VᴓjNzX`x꧂+~>܄[:.^a2: 7(Kzfo0>.HF"zfpٕLC@jH3Wp8g;AJ2)DVn!C)L S`j9p S~DX SC4{)L]߰5אuRC ) aF5BTSε:\fj"헀 a]Mj8U Sx$V .'P1j9)l7dH0%5G=^4TtLRd3KiPE9ӌ3Nȁg; >+zF04=;oOi=vy?H#T-~>y8R;O. ȭEr0_8AF-:_cBJ3`ry&M˛\֫`.FT qTFʀ9-ArL(1H`9 \򔖣j8p]ǫ"uFFΕu㿬.oe|G:d01X1INٹtF0LYRN,1B`(1H# rCxW"f ,I)%>VXOՏ킓$+8o^{sw*E5H_8eBg< ;gp)ZjNYTJk+@ :E=jEkh D|ֵQ!EcлIXFMգ27uUvO蹆/[gmdS;:[O˧FPYI[W8O p>-W4+%\6_.V9/hj[܁ ?*?>dugu:g(/.ԃ=ŚU14)E`B! :1ڻ<+b28s1k2x!)P?y.uBvXH8cJa) @P2z|큕RP | }hLþ+jv,Jc:X-({Gǚc&e[5+"NFHMEr#1SBښQiD;x=,Aj,`=[_R>X$TUcjgPj [<+~h4ך}" `&:FT I:xcTaoƧrrym 쌔Zm`/®cv !T]N6s֚=(0n[]jOcQTæT_R :j\ RJ5sѤt+TOj.$䍋hL?Igv EtrD:=8f^5{^Ոg-5)j $A@O8ެ RpS+ಟ [b?Pd@4S@S>`?Pqc.QM;kN\( ;s  Q/_Y4׻U:I^@v^>KHmҩXtwLLYpy6l9I1S;%SXU-ewH<!1@X>-܂{ Kyu+@CYZ+?iZF%r̦S 3,኷"{og%t )4Yk-zEy^>RjJѽmC0Lӵj{y8˰y # )8H8ul94kqlkӷ XڽfMyԎGTYO;w޴ ز9MAƵ{}xR䦗ixkl]C0cm tt8moB &*䛺BC+axΕ$Ӏ)9%JhQW)Zb-EYr_#8QXR[$ vQ\!#8$HƂh$4%(X91sTxkAp K$9J?څ))JpK~)Gb2})a`Y'CҜHR!=\P{OMT*$Ԭ%d 5(ALZtu2fi_!KN-f < P6]7&|r6wu9;NyRiƵ.zQSnL'#*Y4- IqB fDB *446`aKtBM'IeoJ`{E.՜Ւޥ%J)iRi7Kzj|,21#>7THJCCa`][s8+*ڝ%U~J4[ΜJj2si`YP'd% $u7@_B)LC pSYBl 3K5 NZXf ɠJ$ɜ!,6vgYJxY_YAkFNhj_7חmi! MYQPǙRIJRNj})aQ]ض[CB* =*Tǫ!7%boѥAW[,&tIN}My/ݒݚ6d3v,Akqyo.Ÿә׹zz[u=׹"Ϊ2c#ڌs#_i#Cwšfuu' 'Znq)8"w;cy$h`]Ӄg3O T_Ԭ*AeY QÛsE周/@OubrjqRK,Kl%`1DТVB$,q R pQL9FֹֺJn? QjϮ䰠q%>+,~eقAmd;w8գ]%)4P7*Nu wZMM96ZbE\gB-E{%ni5HԞiPI WL]c2JSR=Ew}*>y-@16oOm$q:Y|)C=aӥ]Ɨ_^ԢSbN&EK[X<w/#Q-J!(֙LO=>ԢC~ra^gDFEvw w`D+J|Px @juB о:J ,)<~Wcrۜ'q%'iqzԍ׍Ď#y#z(g&6&Hz52t/ ]略!,.iT^6QWٌ޳l:$CMUD8NWCD3y[5j׏HvSG|ו{ mA65BMT 70C(!bv5u1%gN'Cƛt7Crş $AٻoP03g-oe 'PX9[ʸI,"%2x<|3ʀQb–"{N?;LL>N=H)|TQ]no.r.\JiREv.p |6άij%Zd!TӔb5Ϝtk3@( I # h)CXǺX O\|A(XƶpȄ8j-B q!Xh @({mtP0<ٹ+G$dT]Sñ0N0c>VcOg1~̫ם&LdáWd4ҏs!q[U*]*1G^}+NK5&\4 Mwx9Sn9Сpq$F?joQ#?<W m,;޺Sy/,`6l/t2/VĪ`ϿfK= 4mQb1G,Q=YvA徥מ!#\߽#^0-HƟ`:q\q.aW\q TQUI;!IiA5gl)%x0섔zK5s"D'mz)`)FORr 'ǤX+t&3iŐP@QJdkHT@eM.xǩRM!#/[R%Rz.Ք~x;RJ)/EԺZ ܖIV+]dwI4t|\3yٕGjFTp)IxmP"_QĠ.d"1g#WN$_jbW$Y^Keg;ur|UnBd4vwZyOl- uӺr&˴_CLѿ =t|C ^1j-m}I ^=:h$R} z;gĴI.]pKѕRto}sh~}_8(iTJ@* * }P"6-Cy!OhKTԥ5@D/]xKo9St{#yu89J5:ӑNX#etp>.C~(*iUrC@=^'*0ͥ5Q9@s厌EmFDɥ6JrX+Vk Q9b52WQQ$IEZpD fB4.Q.Y)5d4M*3@QK0՞]qH^U HnK;̿Qv㤏vIJz~{˪3)RMKgr&1bN7y7>`qOޭZ4'X?:Z,_9Ti ՜]MtkCJO T3Y{=߰%N4ljc4=pטSh/Oq|͞.ےy>~pۓtVd#̓oxM֝;}C[|z/5(}.ZdCK,lF`ڎ'36ՠX NIsm#;s5kR1gE?ei %XkafdX .б|KW،k3/5>7|@A(4t H=x#!,:E-Eoi2M//Ax@-*0x[H D8` ,PRGz(&KNP4 :jODX),n6 Xv?mN _ΚcEӷ|v|&V6 Pd3F A3Ij~kF jeN n@'!4AN X vl0w?&6|F)NiDC=s[R&B&5?u/?R]QgMPip1k$ß$ߏb|֋۲Gx<,S4EE"Ep2(õNS `ճ+OV  |xb&}n[͉-w_mEӥ[Hp( S{;.I:UdeL;T>d1^#S)HP M&TX+4@`aR*T[:q?Sv+#Ӆfܯ܉f(ߠ!B_R>}>B_b6XAs[|η󻛻g%'e~ ~3&bdپr(on~@! *j0QNvGǓ@ƃC Jv5Yt7^+BgȺ: 6DA)?_2,$x%p"\LchXJX/v#k,DWp+F׏P ~ׯЏ^y/Lu2! "ku%L8!.ny؍C!ᤍM5؃Go=Է[m}XJ8#RjBR &RaD4w?ɬJ+t*4Ш/uV߻A> W@\N .lXJ ?V?xhڰ:[Ds 5Ր`R1Ieaki˩ѩuNPTim[ȶׂxRmgwum36 QWF.)A@rdhJE8$#%14H,][JNHP0K'[6j'~ШUd5.F0w{գgT3L[ڋ 3PJR+7pδd*)r W' q Na nIVMY e7T>(Lƽr@P1JN"8!g;qhNŜ?NRK1!2gQc hQ ݾq*UZ .ֽE'~k䩡ΗuvDwlO[n 3B"F:QW+©Mr)H3 `B- "q<9ǫY y&9~–BkzNˆق|4qQ^NSr')λtZur?bs{_~kƧo*#xI]uMjyޠӹ  V%vhkc#}"t cEe?{㶍vݑd6AҴnJf+d6~Ieɲ(Lsx82f3=]6zZC'wa= +~%B>SJ1ZV8̖JntJnt_\*-q0VQ9 VF%p\ +Jh)Rv1RD8TlZǫJ \ 0蟨ly*0 * '""NZJ + ;~n ίF%/_w5gx3$oD9c,"빇q+N`vNmo7}kWWz[ $ܪClҍsq=8:G^2-E+; WZpzjZce QW5.{Ap8<}[j^ȩe!̎ۜgD-3`К.Sd+zw4Mp;jiF{4tk0+?*cGfuekK=FNЍ\M=2_x<}.)!% K}!+[>}d6U~@T|lյ ̽ojVj7P!ڭ*>S2ڭAzS\ڭzvCB"ZI {S1(:!xl"g$7^&>OeDϜ['E5CgT3/]RJ _3y#9sE(|Zdӂ-wd_u|dg\kfiw^ YSZhHdL1Q2[YEeǸaq$anKM&"Ěosj-A񎉳+¹'c5=:͜2l)gw/Ƨ.ZJ B;{NsJ$ -,&q&ªfX'vu\h\@ r4X\ r6PNC!|wBC' D5n3o8u7*p+5ۼsve09|:Fi^q''r)v8wR8%>ѾF'ltfsψp_fÓc6l H)1{~\ot|2mܶw6b!‰mkסLn 9-@*#pVAeNžqP8rnqXjfp0$8?X!"n?) ApX8#J( CPWprzӪ9õy^#Cɚ_XN*|E} (܍X?cdvDE/w}3s󤑺Mv,:-R2sNޣaTT?tzo. 1ݯQ%uT.aIa+QKJ)5Wcʰ au0y*rS f#T{sBR_N[.rl]b6g 3yRKhS3O>ls'<6X}^":UFMAWFAXg5E4~NT?Dgp8N,cn1cCB0ԋе^?dس8J/Rȩ. 0v U6L`@3C<.4Wgz463.KksiW ǼCke #22j (QNߛ`)IǁHi0,qBa.`R*ŬٜZBFvɦ;TiW k5f92mG<9[愐 Q\+-q톬9kI:sT!S,f9>Jw1`rM&qpQTX!,ʹh= }#is6hFoe=&jwZXĨv0N(vQHAx7;X mçU!e I&^lYySadĮ< 򓄭_k^.E"3ư]+CDϋ98GHLŐh_1Q-ip*RbŁx``$9/1aXF]Ǫ7|eE?aF Q@PP @ic*j(VXRT6WA%ILhL/XgiᮑQ5B&F@ ѷP^fR !tY 㢕vnm'- !F@./mbRy|#^X)Ě 8 HpcS'0*ճe3(F9̫^h,6c!DKH &4f!`ZW͈B V$͙WAl=6ݻ$PAqj?nl +4-={j_@yͬ[AYSߡ2#CPU"1Ovƒݶ@,8(.n h[]);Nza;vd@]q;VvOS vșZ$dV ڭw:>NDH>akcwo+RA y2#y+6K Zv[m{*s,_$mOFƏOU7%;`Gا.iB #OQǥ+֜Gy~4Ks'``"DcI]VlU]csNvfBh&BcZbyUdފ3)N޶e?aE+rޙt߶wh2`fصoL{vӖƭ] ˍ;hD3\z "b _W^BBWQ9P3B%x=y䠐Z;nW OjWC$h{ÇlVnqLU9#w>$Q;Թ=6 56idiw:vYv^mD=e'Cݍ&.P`ϲ;fEL>Zf&7Wo=|n>ՎL9|ty^wv4e/}>Sݾkt<{`*dFW[ BmJ7MzJ8n8vTI4- G!ۓ ٯw_Gwv?y3n B,AR7B߸?j&wR0@c҃зǶnLI6 ':<ⳣ_'4ȞF RSאY&FRg_ۙ?ؽ%Oh&*3LzK]X8qD>SQ]}o"㽫FmZ;'NCl!broԡzml=}7;YV^f [K#!dVῒd4kbe(+,}W;\< @U!7Tn]Uɵ53, Q߯13#R|TS ,L@&5x_o]!Ө4ȝ@iyujB`El 3;.< 7 HBmD+k.(& MwBȁ ZwhCgr Xs)BI%_XBPT^:*Ļ(2=O#Zi Q u2K.[.m1^#0(1Rm>mBe$6 mXU2ej$俊 '\m܈g /׃k_LEnv{iٰDWvc?n41.{{15fgbc&1V LGu"WlPh3&wE:jfsfHJx鮘W+%ϒZT%׸_L' '0nz:Ӈ˛-TOgzFіb6ly|fri(˳U_W/"SB]~nwSLcfJYwШEA̝sink[ =n&&`,Ah{ Z(m4?TM"1l=o䳎8n6{gaT#=QTF!=F'V2"m!v7o?L,pW3O5fcu!oTshLhUn}b?Gs}t(2epqv>W_ WCQQ:wbQ~E1P8i>w UV7e#YD <5w A }Gv+݆encXDٔ8x7A;cnCec:CŻr!$ϻ /ư߹)N'y_W ~ Ut;o??z$\tBoV|u~Y QJi#5[unyB0OUhdӏGSM-_O|~lc 417kkhG9}~B˯-tps]^* tTR6 C:R0A- F- AJQJQ"~6`VR2}<eܑ J&P0dT9x>{}F"M\տ~F5pc ^bWY~:C nԒԻDk&W}بE=\)S[&EuLIqfd*/Mc}J ll65ص ̓j{sQ?W PmH{I%bqWwC};lZo_b~IxNAKѡB@w\2K/jt뼷vTFi9I5zj^'KLN¡@4 +NB VaM=.*}Hhr79/E҇MA)RSz#(d#Þx@U(TKL1HO㰧'l='m(;Gw2W‰QOU䁫*îߌo0y?۪hb d__}\W*Tqh˜GEJP5Al6 -! dXk`Mm*8㰺qqWwCz[>[-&2DF/_^]ylL5DgqGG%66Dlk ޺PL4d zPjs7s( VfsWH1kgSI>AYH$ 㥱H6S*`A# UJBIZ:LjR'S0&T(ԩDy05x+ #tEZLvd{<3Y /@/GyQ߷ ͈aU3zwV,U6'9X..idHdW_N{|v}sX?VYE9SH Xip N dLd/e:@Xrh "JI\!=r'ʦ +>Bӂ;񾯓Q:yqno6Ng[t[x4O1wԜd7ܧ钶Qv9үFݬ71J~'jJ\]V5ڻ_ڪ*njEp4$T$B0JljkNN nsmC(Ew$KnRPWbD&0 * 5+GEM&PV2"۞2 T)\5'[ʡo |meto"B1Tų"hegk(q?.BuPBTkCkQc*@UiP`:Y\w=VwO(=vq?!E,Hj(YJl%#  %hSdFL*,)2<$S*P"4{r> /u陮0vj(ѹ TW!+{~ǓPZ vA 1buX)@>`twdvwՙ\xq9y)Jß9 {- ^1AcLJ5x%Z Ν %]+hxMmœ{, H9 eF ]*y*O^1ȔS 92Eh3.2lcjLcnzd$QW}F6ՈBY[4qV*Mi81߶FdJUѯ(1߶Vxgg6T9BbYOHfK1y: {)E΂22ΌL +`(6*jnl7mgY6+H(&HZ ! P`~OOO44zL/yںE]^yL)%OJ3(O[J*$VR*6.hdLHQ$Q1Rl`",$ ɒ$QÄRp$_jN(XcBĪHjLJ# @8\\ۋV;S,ͨq`&)Ҍ%Md$:b5pDܩe&+>mV^[Iߌ!Dk#jEvpeow7pL#>.CfFzOl/u# \@q·bdMK> _K|gZ>d׸yKAi!6T.&pziMU-&E*C9#eHUpm瘝C)\)0ۓVY!y)9l'"٨E6_4Fz75l 918Z/ =):ԞU`%읓jy&n {jxA}QeddZy}{٤#Z-\v}HZ1YɱU7GVf?Mz3$-x ==";_{&6%;S@O33hkbo!tkG;$O9,˃Kv_/'nw m)VAкagnu8lԼ>Z`qO)P$}ZRU _ ^);akq)vK-mH~ƘڱUP/!FU~=.9 K婸CcJssXRA-xu>AXN|NoTJk9,N'3JעvUH"p%'e@kE_fyf6+e Zф lmb4 4qBeph$1Ztx%7>}cאַ7G۝?^7m-y.up'1BAm7jkpd12R{0u4uqQt5ܫ{R+|ў|- %QRD"El.viw6* 9W>Mi9LcuLwi1>.͂<cKaĨSH!%>*Sm%%VqFfRbk2( QiDq%D[M 92N \b% ք~TEIJx bU܊QPʸJ#$UJ,2L.qw*LZ޻`R/{?E*`o7q>jUDgx%#ӛψsr Сs~1٣u38;Ѡē(Lo(4+ Mt}y.߹1 ,Ty8Aa{ L%B!9`E86XG(A6  F[X)Ehք2]^S  s9֋@SPDCju4pn)w|D?v0x-LT<>֘!ĥDNy*8pFD)AVy-LL8hړ-+~M&##F1BTt,ǨMYs`<ߓ(baU@ݢ Yb*Zr8Xd GK'cYǁ (á}'(6Pe0M*y/0DˣVlL0ŲQ<8AR^|M)Rk?myB$PO47[JZ&"k!- {.`B\rJ`AYH"B)Eq%%8d07@>4"%Q ,[Cɼ-ŞQL[ʯb Dc)v3!nRZH3nNbeeĨFcΓ$&`NDT Xg6YUB[ q>m @HpdI,NaLRuv[ D# H]LHQ2@!y64'l |" !XcNƞAڭ-b/aH]=xRi.,;urSZ Bv=1vTe}>˽΋G]FLc6xWyA\]ӵ(NImwR u,$! 8< #o$R;RP֧15db }?^Ç?:ŏ}v [/>uR:[=aXOWC{i/0$dX%a,0J,BhcCV%$ "[4ZUջ57$2ŲxL/P#$iS"cLC+ Pѕ8<".#V 5FL0$q4!r2rUj30<jAvR[&eu QqM)CD0e,6IT(RbSr&FF ETp}wgW@| CobD=Hlm^`깄o̥:w8 Nb8k0:4kS;4O#g͟|;7mApZ(=t7#2 „+cc,JZbtSV#y40GP$!82t`28^Ɩ3B1N'],đ% k7SO+N>oѕ;D4B2>7x~wF~[A?{ Hn7?]06IۏQ'If^8&"m&-[4J; i͛qwqI@ #Dw{c9!)a3No54Ln0r޸zVuoKO@|-g7__m'=` >;k<{x[wHc4t߹6T6͏fvce:^6-?CM>,4moϼ ׿~ cww_ O?ͩ!p7>in.3\LWmf{O5X`O{n3p^_mPHO'^cs~~ٹY>0vU΍fr?òИ{t;;5p416n{ɛq]"rfO kiN̑_wi= e8,p3$& H1K15(eJ,#10gFKQc͓G2xabjOˠVDc)n+XUiD^sFf9%>idN$:NN*YYxAȊ5G^܃>]Y])تsSc BZs@*^0F 6h & c bD.T"0S@=<[]s7rWr\v !*u˾=)|Ab&]R쒜ݥxgI4~C.3r&$QblV;_B}V!?J>jx^J#46 e[ V78fůyh;\N ":2ا1Hޠ=p/;얿\.ō>= Մ?s]￿$]v ~zՖV/7o ^[zdk>M3ȱ( 0¾ЖefKN% s3.+3=Əf#GbUg+|;ޏ~Ê-w?/?{GZ]zofhssqw9|*Wi!] + EaI!163^:tD"Nw 6PFNXIHNݝj5ڱO/ֳܻҫ; hPLQv[󆖱qG-aPw9W* E)dt|8˂^ OV96T0ۼ D4[ؕò0u*]uIu&/݊uh̊u9K,F%cL` ;LIΕl|7wodѣ9E9m^sN9̕) ϱkk:,Bd'/(P:<%8*K>իTX 0tFWBڤ"`&Z[q]v^Ҹ&+yw8m)U' Z'BJ&h쫞rRO%BJ&W)=I)uRf8ryo?KK_*Vѓa=Vp"?3j8DҀUXps4Z Ibpof{V7 %}vJ@P4Zb&#(AL[sUQm DWzXsYC챶ĥKa쾽SO.$1 ŊSF=&Cdzv!22pQ[Pm:_ -eÎ>nGzF#;?V_Bmؘ'mf J}pl֤bJ1#/aK%zrC@}4}D/~;5z7y0UO|,d]N˛1YS}vƎ81)%eR 7B۩VMm%ԴX*:fkï1}egFft?Zq x ifZ) #}x{uR m?C^%.8օH r3B'cIzG OcY! ʌ?dXz+(XL0>GJ!#u9Ӗ>@/7Pt\Ul0!U+Sr /V*-'Y{N겨:ͫmK?_/MWo2)P;r>s &-#FSVB9xv>*X+6–dކ~K b7e]EʺbYwQ>'+Jl}!kJ@ aNowF92 $c?M?:V#Vgqk6#gb=`6@hȕ"32^[# ©εXS;d6|BovY}^[kC|ى{-fsdV^3Qljmw cP7QѬtRR#Y^ϳ48g^dnTqhN?|#ȿ@\  pϝU(Bqvg&*A3NqXZSt(iUz,?8JX\6vaK ;\PxmјEշ%V5BQRߖMV3\WC#ȫ~zt&c"xP+NkƼeo^Oշż-aLX2zl 44R??a l .߻=ZZ欧WUBRýw쿋`]i5EHͩ-1ǢtVyAJ)sW "#f9\^Ojy|}?t>8ig˼ O6^?Y>݊5{PzTz*ئ4zG7&&ūn}ģH6tz52N9mj-Ktz'u,;,!5k@0u1@9:x zK+ҿ} keSKz]LK>=zrk9mvx(I=X V q 5EW5ݣ{y;`Ycz? RU\4 VK86"=dL21rŸ{&_&_&_&_49k8VY*Dl*ԠU% IlmWUz~|1b9KQ^w:n9iR tAʻ\VJP.GtէDcU>%=\v oGjw~$MI^՜%p#%>5])Kѱt9|or[v :_tUd UQЊ=g*I|??|dբ~7z.y[f?}z{&Pf?훳/nWỻ~f`ཀۓI0\Hk-=\e?OHZ13fw]cqN4v>5l責 f6zP9!|:sE08Q5u]"|MY:Pa'vUfe V 2J栠@e U_)tzxgiFQt!Hu9 feVhz!Br!;ިJ٬T"aۤEc8w~m4_n,L ލ̿L/?}⿭>NV s~5㾙ga;_6z8~?A}V{@WΟL3/̓c,RlʊkXf!=/J[#[wyG7k|4GHixn)kJ;Mvw{3-s6”%VRb >t!o?ޮip"Wck[,7WAMٛƿOi͊y?Z:5S"kti a/Q^Ѯ=(6 ).y?Lht-S6-VO.ͧ@uZz|͆05y1 JJ*֨uUVlERքPRjp%dm9i8v~|5ƺ& ~{{ 4[vBNؑ2e^Onh ¾}d]k=>q R}a1?an_vDfC;ڣ|zt.&c(fG YtG6nV^n>7:7 GPH\%z-FW ( K".729}ɰ!4t6kBVc՝j@sbK սg$tP+w`~uT[q.# 2*P3]C߅L|!ӽ%DR}]v~9vswcs4>RcA + ePѲFWyjt"ysa$o0^h7O%/G A(.ӧ~I5b#%2=4IJ JնD%ԕ]I!Y#X=a%*bF5LOf1ܻN8Fh%`DB&͈iiwBQ­PO9t4ӣsq79$pZRB"~k>APf\Q}8{&Œ?gÈztUZFr)Ӏo4uLq%YW6ޑg([J<44z.#ZLy\zek=ՃGm66zDbG'zYh> 0yKRJ8M{_ -]&CѰ^y*9lTrH J >uYu|?޷@~(3v}YnR{-VflS;ջݲ)Ue""+PIhvSZ;ώuce‰g 8g}`svHbcB+ vGxIbXT~HLoHZ֘&#Þ|g!d1捰y1;~}y-2PdJ*>p,J -?]$H j$._מlQ؛e )6&OZ{2,lo;= Sj%E.uYwgm7_=Ɛgv^bB#L-HV@`gx 3mϤ2_xYїլŵF͏+I«1$)'߽*tn,6F)֛wqۭX ^"m)ޔ0dwg${UJi~CKYxjTX|a,=zٻ#D.Y |$lƣ%6uRb쬳y><jÙZm4'aڶ^eBMCa {RDPй{_0fJd}0Ip+ -2˳)zi+P x 3y Z˚2w9T,وnjwn7E,ˮ6`s[JR^KKif)/=)vP'eoNH'X+`3%TݱuC'6ur{7+\' >{Vee5"xo=eՖgɇf$%pt^}[lZd ~XEl5 ZU!ٓJ#㍝n迪a,Āds;n;QrdA-G8R43Ġ.V[9'0a5?Wt8 ˨ۿh-ͽ%fЂA^(VTSR:M 4^CܧaɈuDP\RWY0ӚAG}H ] '!+QSmv SKyܽa,="X۶f)FI"ab5m /Z"{ʽ}'z i'L׮е4H)toz MΓaA-l[2MDג!{4-n ߮{_.%ȢXٺbɁ CKwDc{JM,6cG˥B[Udwg>75q_"ʠݏP Y1>4p$P944xˢJXv1oWc1P4HIl-SJ#<X 6Å_>f[;}絀A QP BiKp-$GB*#Ȭ hݢWKVݷ#DVLM -R-m{5́gZIpv#h< &/8/ڥƉOM%(@@f!?jza+.iI<jb]ϣȎ DC .g/WIcmR˕?YX/=˙+T\D^T"!lֵh@goj0d堐7d)ڇuu̩+-ldOxfO b]p{Wh4G5uIP DSl`Zwb?gjZ4S2zCbE5̄͑hO~Xd&toKeKwˇMOlU`h$;ъPذZ};j˚wu:{[eD98^<TBND=Q$!#oչщ-23ޫɭ0[~-8y4\;AI&V<^ C`;sbMOP±2!"H rloOڿ0%8iL~xҎLJ(`N*sQV'`gb̆BհZa%:I hyޡfɲ=J -C^bdޔ:<f 1+'6Ӥ(sEZdx r g Pʳ`|Sj,;{NN;s3:}c5X;u.jkvc QTB %vk'xwKq\}[tuv=T]I̮:vOW_Mj,RNnXǐU -`I`ćBLR:BxjJu%P̛p;Δ&MفpBV:ᆋ?]_jvw!OfauMRBwXζ3)8_|Y%+E+BRYSђBJ6npXQ\(4ĵ)a=^,aM3Ŏ #e GM7lJe>jCK V5:0k@%vɐ|k:Mi DFl)Z:rocY쭝֦]hK2'WQjngSʢsru4Z#HQZ𹔢jT}wTţ{PٕuĮ͑z(+%D&(^l t>]~> )\Q-c+? %S UZ>Sw9IRXw݅żGŴ1-ހ@QXOPﳺv_`?JwG+Brk*Gi/\|P}(|[!`-^̙0bRBŶ|FLt=%.Iz[f`n tКi˛ɽ!hbbN$*6 &PH2-NUX%+TlK#xa(($RqD7і~nX.FwR45zt9# gt3  d;I1gPPwvc{ݚD.ŲYV닜OSb#gw՚{4 #+=4h&W{JmNN}xٽ}olǧیs1?7d˿eߛµ/xV_[֘Dz}N]_0;8QbKq"9_KzP&#'Q0)$& x@ϝsR,oWR*!w^o@+վz :d^y;`?SW2 -OEr2d(Kg{/R?QW"i +7؁߾-N;vIv$lQC Ej"egc;iz 3(L`ȗI1b Ɣ%gBv@.Ck,^hFk:Enx1UȟE*7x\ | WR"rt"9U7 )munWOŢG77"6 Ԙ EEwE?}~,c4ĹZKu98[g|7,m_-+`'LQ.94De# AYsZ#])Q<{j32Q'yn Ol)gQOS/vkTNb}zICńrÜ\(BY[ѰwcrJ*àxKdt0Ab"xy4ؓ#5݉Fx$8ʿ1ב0~a+y$D@$6<2ٍ '7wWT:¸N$r2 jp3aHEɩX5ŃЯ4`8:h9ObxAIάoӾ;_||Y|ۋ sULs]Z?{7X1@L:'4` ^.JR;DDMr3 ty[Oy-& w@+1M7$) aYĘ>ŤsgqmKmE_ 6%*T-㴼w]X$]لfK7>R+ z g@ .[&G@R\&<:s !F+YoE}ց|E\r1bddւBdkuP\_nQނl_V>Ţx3~F`hcуտ~hr:jfKeݯ!j>{%=dX}P,Z7''jߤԗfپ=9P 'Έ\MFnIsk&$ ޥDF>]3C|/#+%'GeMyQVF5AP_8H5%g$e1,Vr15 =gGU);iaHC&忏 ~}|PH0Qѫ͓=58g+ a6PEΞqҊQMg.EC+?a~3J0kLEKωxU!&;L/˶/>7ofR 0- =9HssgJ+GFFlV fբf_~.5s'G#ݓkϸ!F,)zm 4 *a.>xt#)apFIz(Q(Iw$\Ah}&xM$sÎHN0AIYiIL E?4dTڪl؝{IKݑ%/2g_,9(r ڕqbָ7@˖H٤\u޿b*fA3/y3d``zq{J_s?sy ℞uv2]F[`AwdgEM^.˭&.+{[2;azI#v9s ip E֫onI a;v>m2OF'AI/=a|iq2A1D3*",v4&KvQhtVIBơb3&#'beS=>߶VIMqJXNe4"NKL 1 2Es7=#Q\l1f}_l9%7/MEwvރi;1݁fٚ_>:??9cDn\$2U% S{H~дQT|xQY arJOa?w if^"%!'+~h $ ?,rEu~w0~}/nU[v'v0iUGD[D9Ӓ`_"5TG܄׼0 aS@(ĽeT!jJdg|bAJ.SOQOacR#\ˈ@L21O"@5s셯!`Xd+ ×.1.h ܧW2@;ca0S  aeEEx(/" W`kĈ0TF\t y|9*"8V&_ֻe&}pW8fU_f`/jR|7j?#5gu?%+_u~G'y| _,3}~SpqRN3<{J:S??jĄYv>CX4ͪ-LK[zaN2f\-0h"3_嗯k>0iҐ JLeJtUNT8zT\_ZBa[їo-vXV3[jki(:+M+P/M$k=?X$J?IaI=m˽̍X{l6t6/v f2b# MǧgNbEq '+Io~dޢ%їi'*PIVfE]x@vxE[fhyռ[_:Xs'ٱpzj gy1rDv+fPU}^ S"8jkE W,ǒu &UWô̷Wn:ٕdLLKM&!)9Ш ݍ9*aAOZB>}]=Ǻ;Fpw$˶xKsyǥ՛e[4c+w:Yv׸ut҉V?fgVd/I p h>;زU|^ͭt2 _a]ic^Rlkd*l`Я*ZgǠ!:\󜵇D!tr<{q8a @[`%CcNA~!r)C48D7Hy~g.XEJKԟ C!CuÇz!v )ٔИ¦hw1,-gZ*d'>q4p[F \aWinX t  O& 9SvżU$EӼn+OY YlXvBNwߝD2u,;n0w+8F¶ $`}C|7Krh\m&Du=_ؿ^'ϿQ 1M翲p6|znAҠZiP4hZi}\ Sm)1.݌jB9^3_b,P? ܫeD+ q,5l<:ہHj;8ĕVLު? 'wG5$: ֠s΀g:U؋D\ Ix(FH, h\qOdEiJ+]PFŪ$zIy!^/-ݔs Ftg~8 ltO6PmtȆ>Ļ駒+!R,@;UN(F=N 80Ʉ})'" B150&P%$QxWL7ŅJ[ .KY\ח)k]$>m> e#r?KA_rܣ[pii{c*5yrҮ|j0z;աa Z4u-n{TٓN*Q 7~dKnÑ7>d=<޼"{f3v9Ώ?,TPۀ0aV 2mb|fD]>\MDT)"Ά80%KA1b A2:,&X)r*zȯXR_,jkXc8d,?*2̣&%SRm~]lX@f@~+GGA(Ӆ\J=f'XŅo7{3l TaGB2$i[>w(< :U4ƃ𐍋8 B0%j42}ppPƐ-+"zYC .[nYM@j/DWf;Aqz6ܜcRO۲*W8;eչM\q瞦Dp v&dR),87*F@pzKiy+F9tI5h-̅(YP BB5!AmKp%B$0@^zDDJXn\8$a( !mpޡV~bYSdA ǴO%7 ţ[#2xh`InU73EbC@Pw75IxNx.U6E[mtRr~nWUmhT wٗ''gf)|vd FIL&\;IJ}'*ݒu[~h5:Ŷ \'xM ;w7G8*;@hܔY _l22i/3CoڸDpݓ6EovB;9 & x}tL (~X9w Q2q).{SOYgU{!FϢj㣟iP$*0Ȯ6J#'M F/|xU>ҏ_ipUE`S*qŔlE,͎+qm11D`"/%8͆dmK޷ Ҕlך/6FT\޴araN+fKw8)-A]A{M;UiU6}H|4hN &=QDy-}e̵ȁlm-j`ˋC88D/Dx.+a P^HJڡAQJ:C]^/A P> VF]be0gU_\3V?wµߙY؜YE'߽=D@ U7<ӯ'U%0XI~}s+Ǔligs]u!_|Fl]|w k1Ao觳WZFwϫ}(@KG/l4gU`G |88ʩޗhsg`UUާę2v+^u5^])#[ں$B_h+I籌E,C>!M{(JhX֦5NAb."@ߝr%Q,9f?y;2ZDbY,i 7qw7ht5{?<;:]g +?O;nQ-n\"^6w]6{4X:8M-|BM_6E~^8'gHlRqEl1HsA^+ƳńSQC5 IW“m5Ęo:-ؾ}t߾׻$٧w n@lUeo4p];X8<\U Q\q<98{鸱 KaA5$$0Vmp+K\D˖ ꤗn&W:|%MֽؔGjMRz#q>[%g_><>Nf)Ś|q6;}ux($ux쇳U9UYmm}t l*3̃i#~ql;hp-m;v(rt\VXVik= MǶD4ε9$5iD6)&& rBzi"]L1QbžGt=F(36{.[g #i؁U]NŅiFaw5801x> T76[=>?4/K 7"5O֫N_-JoܮVS6JPnYiM6M#n: W *ESU4Gjϗ7_ArY^]6]pdJ.=\*F WZR+J \6_*Qy~3ͧba 2 ^".êzlUZqS ~yo!V'H901Uh5I8R9%y)VEQօ)_r_,Xi 0Fsc(6 i++dG°M0D*"(ХLQau^Ж+)3 2ah^0mCB]=$#hb/fAbڌ|dg՟}"^k61vjmT "TLf㎇iɘU do%mTjwVw :˸F?0l:oSd QxA9h^%ZWƀ!/w*-ck##mIs\"!׺$1)2|DDJpc4xGDhwC)PQ8dqPتH&kU҉B_x{!hhGi*5\5oM#`OĚx Wm۫mӜu^oviwݙs'U2[w$A-ƽ^^5 o)W+\5Skށ ܁^=9w1Fn .8zIM̓368Y,)!)!\K=NӉ c7Hj4^x3 〔z{{|( !18 $y8 9F9{i't)O5 "WsRy 7yt!zx0vl'0'f_g:*mc׎V+=?ɕiN98صDLvcג"LcR:}5c9})q樛<|0,QĂ4P-rV%dCL@z@}jtNvRT`d{ L" enBվ\QG"1V(Qt-*D6H-z[T$nQlՃj Yj}&- lfKcv|Nk&<:6Z y<ѳEp#mpZi(73@ш,Wu?>\ɇzӾLNsi cz`6P9g|A.1XFŻ>PGofHyA8 u. kPt#{) xWܭ{lMmτ.p8%v,,%,Q%rYlO - 9SG,@g`i, 23T zとј1 FMccs!Dvd Faԍ,TI8`Zq_bϫoYqx]/d'{5[PٌLgWwLQn*5(.uL`8qy˳6C&>+GLZNr /K!hx;g9Qޗ#.8ω{v"x |붨fUO'*$bdu @r.u˜rqŬ*],.t jNiP󘗄OC#pJ$9^4Ykxn #Mh>8 _>efjP? U<\_5" gd6olEd+jlf;xRX~[͕P1PfvħF4F "=ɡ?<;/; 3;he7u߾ۍ)B6vdq!6@ [*O Fptu8H >Q8'43f#>lM8~*mf G|aX!`9щ'R*yN:Oāρ+7sv:'=^Â1NkX:88 Ҕen+9t:W$A@N@m-BoPѾ5]Iػ&ekhl} %cp4|41砍"I)-],ei5YV"|sME"p)H 2*! 'h"%\0k2FZ!RJ-su#2b"qnѨc)WjH@ ~Jmt1JU47H2v#G|A8I% @΄#eNFY<3dhkY2K[tm"aw U5}lk%il+8)^ dW M%a8ۊAͣy_ٮ9q Zyxg3nģ"·Ā3 r(}b'pWm). gp/To3Gy./}A9%o_v>ҬkOK{zmrQBz ]0sFA=D45Qp6b>w'0g<\x.ɦ?{+ub͢Y߲^ii8؂0P3ѳ6:L > nl kMb6;`\u;BP#-`=1%cDX)YO*=EҚF77F=!V[AVIxx( qK„c+䮃ɹQv("k {y^79L#Q1zI.Y:Gę :(A𡍓,34ൎ БOk톉٘rA$c_QvWn"?8jΔ4bĩ%2}[HNjA-U qii6SROǫ>9'N{:,L&-H>Uj)=?)]9p2Hݒ.|a-%֘)bcš HHnd&G=Hx `};?UGt)kJ<Yal-FD&gѤ! `EbQQԌ#nQQClѳޢhC5o9 )J jeJyK)&g4od$4mAN-ҤT5qNCJ K9K)$>y<:I(o2y1Bd[Rs(IxA6::"6Ŏ ]CM.vy-nҤD)P{]Oj7'!iRZAMqrsRTiR]NniASygmH,pC%Xnw%@ʒ-QAU)i{dBJzבK# ǥL6[EꇻY1^~=Õ-nj* i ӿ6nmuԚ8|9K`;0A-kM`m>]jtooyip|h?YVp#q^~ڔ\~}7xנq^)D}A0 e,qZ+6>q qŜ7R!ya+yE~HE-84"L")AC ɢ("G֔.-=Ҡ R^0GɔA FP+g@)*~[ -Xƛr}fMԕb k$\Fiь%3O]Y]5(yQ@%SZK"AQK<ʆX*sJϽ# q#5{ HĿv\Cp _ Ʋ,Q@I JKI4 Vs#D$*FA :5g&K-l]%[ZT_BNls79szHC f)B+9M3\k~j(Zc h ?0w >|TVp^%C&*Vk8U%0b{tw5u4gLuOI&#۝H#%'f 7:@x j9JG/ջ 9JTPD M̡z*Z.۬94ͱYY9W,VGFꢒ+Ș06Xo)9f6BQT1xQ:f&\R[Vd䒗E'(_fXvA!q5n{@S4A=Wl'j?@~>Wl*D. ₂r*2kAP'h732[BՒkV }S ~&z7o4{lp ),@jj$ꤞ`>p>um|wALd$?FhĉΘqћrM+ڀbrCCh&1#eZ33Zw{W&fgd^|]JN"g{vTrwT{j3=)|3,60Ept}pD6KX sSVR2;3k ,^6%Ox$.''ƒ+\((mc(C)?W+98v< (RjMA.$R”-Y[D3#E 3(ɫh^\&͛WHDnocsm&㙮9t׫U81;z|fV^d_>فh=cre͐X?D&gs:j࿹}:ܮD :/h2bqZh^RGuKg,xSZ9M]Vc e,K}l +nR4yoi]|f_WRFw% Ĺ4+!Λ^K=H Vٖ/Eoa(jLqʅsL)\Jٜ[db\F{,wqr}}I.)zqAA•By3J|ۛ wxII 'w˛Wi#sfhKsr[dRj %._ޮ̕^Q("\qqڔ#V1`&(`;+32pwM,a!RcվFg FX)`%PSjRHM&T7`$(ao\SXmBJUTy!u4JOqڦ\Œ$hXn( 5BsSLz];jT@[;RⲖS9S4AO9jFz97 1a=m$%RnV`ѻ(E@ hfoJJ4 eOjcBmA_}0Y6CS(:P&z5~<[qU/{̀W~!yAzoc5)"86zZcg&A Wsm@=ƜtAt\%QՊcRZCHu\Bԋo,w;^7kxASoxUBh&C4ZM%,`DƋȺ+L3Q#x!B0V3c&4#4wTݴh)ZGgJ)?Jic*Q`G"U-dF |V-yߥϋ+j|d'6J $_n~'r6n"jGɣ \E?,O^ mZ~eJjz%ųH{:,UF/ze//UExJm ޛ!:4EJگ8\W-fIc7Ig|}f+X,[f|R6JP+~߉w` #vVԼQ.C힚y )1t=:DU]SۊiҏO,RnEjtR_T:ÎwC7JޞON`ބ(A] imʎ; %N qhL8DgҷK_ *8+gN gV5rpw,lMVi3Tʘ%}׀5B!J Qՠ=;Ҝ^5zFKl)BJsH m85]m-gKZF+XQEFuZn `\}\IuԒGKaArH?Tzk ΋d@lUh"zeT<0 < 8˽#u8Ĕ^*iV91Q{ vV!7gݝ>,v* evOZAoܝ b8۝ZUzHIK%}?9R=iPQ4g[ +3O-̬Y[2-۴lFs:qYE-8aHmA[[m@Q*ksOkqb3=Pφt;ʄ9[2gO{{/:2>+GcqGn;qEE]f؛-rj5s`޹)_Bi"kvQX[z:*jM(Α,X2sN᭭V^Z}c~LպUg>R{qB孄VEt-uk3ӨHR&%b*JxA8U՘:|~`Ufj"/˾jg<;S h@T ıZ{YvL(PcXk[9՘#r,N P K5 /KK|}PIg s(DluHo7yyc&~y4eiU:f['2T%cp(|)$!Yk-?.V7 7[s! xȷu^G(6Ɲ ATj-؁0`Ē:U⣝{tXlZQ}l牷g^~s V)1?R&B%'&Lor=[yneySԬ51$$ [e[ IoJOvF$ 36}1vƁhՇk~jM\m6<5n@P058`n=&95(.ӛ>xX]$=y1&|aH(mC~;[Shu¢"jDͮAɏc`c/(۹r@}l b(Z84BRH AN>ڏKwĥ/$*=/ؔ㝷-n_]Y4T6a.drftjriSE)cI>j(N9z^ǖ .;D,VOW8 &)U>% ZH^L31N2lc& ٜltbg4:ʈE Ulx!J!=Xv X'h&ww8EZI1 {.L+K7hCXhCf|lUZ@8(Wҋ$9/W'{5m)*1QX=G^L1*.ݵߡ ڏc&Ě i>ʻ&j> w #ѱ E S3 /DAkYSLC `Df.PTsF_!NKI{h[K R- p;ؙUi;6ݶCa''ymK]F+zJh׃RNC)F)GցW}T^@%8Vu 0Uϲ2a~d#vf%E۰ W[7O]:M,>{VhR3Y/"Xjަe" ǒź`ÍiMF@eM\(|,8eZ ɩI߹Ǵ<5lg$c\<[SVvI)/ dDڱ BFكJgه4ӎBYs K (Ox6栶H+3mEZ^ԃ&i&SbkDRsxp@mz`Q((k8l" fUQyNic<$oy'=TVvxu!Hlhf-[Nޓ類H!w4coʾ%GB{$A BVDRM8⪥V6-*<0`q'FA?OMas%M.D6VѲlg(1'fJì{yL}SjivWN:ffmd3ÌV>P:21P8Iyх6ywwSTJ l^by|et"^c!|- JYZLuͩʂߞ?u㕧]2Ve_qNx4 Y +M(i?ZYI8Zh)+8eiE .sx2 8;v49GCGNVJ`lZ)qXK&4XM~8,΃;T8zA4&3y_WQ,`qACхg`$&NH Zc!08r?6PJ6K9{y4[MvZ)(Baki)JsW<-pm~xX|S2Pr2 HD2Y`Iub{JΓwh$ngQ<7*EGs2:V5>9뇫ܪ,3Rۂ*0"Ö0[J&X15 {}oQ|ۈxT?< ]ɳ-V "GB&JVI)RÃSrtʫU;#sX_T Ep2"E[ZM> ٳgtC``|I~ǸnGҠP1FvPGZĞGg11jQCqVcdTAD$Ɋ5]NMlo,EKj-rZ] *;N 6?vf<6tޏvܯnt[_b bFQJubz<7cc@LZ3 ZsIad>9CSN 1^on^7}_R`Jv#jvpjV8YR5ۏYi.2zgl#fݱA JJ;a"ԈBD**Gf֎ClU>uײPjZF%<)6a$ ?'s9dPnkQƁ< %֡y1r,5/=;gE7{" Ta)+l[1!_>3 I@ƦM J,!zjIO^ T5$^]*I#N%N]uͭ{|^j)Qm*\}PiBvQh^_q3t󴤩>SIUnR8&L5!~)1 ԆȦ tW6xjJ n~VXW茓cX\Scx9|INhMbk*b}ɢJ)ITM" 2f EKRj7pZkZZCؕc2/A{wB|+5v~60y 񖙳Hel5~/![vAXJ^-V5 C JA6R$^^HJȂ<?q6OCxP݋uwµVRʑ#M]FlJxi$pEԲmN{D$ ;nz?p?y i4lD !ԈF{f7d "_ʬLy$FkK-G?=l0R}$ÇU)'joG%hH1gFmX^Ya fEԱ$R(8tȦD 1F"޲\ +֝ _~}}"XbKG "E MGY3 ݫǵ,Aj~UUf xˀ49Ȃ'!ǸQb;?#Ls|0lzk:FRq#?M'8g^=fDtz{>vh߮CkpmY1㩚Gde<׶4!'U>7vDQuɒP yiӛwDzc(Kz&~Qa2P׺4/52<ΠƆc$( ,,bFˤWMSk2]ޛg;b|Ywj0`#r/A])rz`UnbjQO ?_IB*N?;Owl~`$.'A~ssOr>-wR}U\rC9{{䒶HpTs=߱5kϛ|]5US86ly1x(q,tm=%H'ͳÿ-kC WWdl,/?s`ί1`D$s/: L{|`7]W= Syxo LLTh7U:]Oo/Y#QRZQY)AjnHp8P:,ȎEsZy};oZaZ3vMGB{DzŵPsn0,/.n@/7nQh^CMdIF0)΍sZ;]rl]LCfR.t}@LxwV4p z,'FnI܇\2QcI?_B~Y\:Y˼<xfsGQTTDy/uNXORU5}[t~WW+VK)bo`]LN`$ Ճcۛ.bZ.n6Fg>maRUtCN=-ϩq^_<1|3^Phke6l*,<Œ[_/$Eᨀo]iCwv}$$ήDa|!6jH. ʲI>2{ P#'*JIrhw Ò@[Z BNvt[[[gӓlt}:f΢A{,0QLw~4$-S ټ_6IUA<مΓ͛(g먍VbrȘhEW {T`N}NR=4Gp)2 |dJh w}\ G 4HK-a5ȦQ O 6:Zc}Αpr(ȃ L )Fycp4J:{N!vVSØv|1 v; |qWMFɠu"-wI'A8py/Fq>`!;8?~'- yϗ& f8V& 3QYu"vw  o|+RZucI$1ye 6,[C(%ؗrI匳(})x:B@bOC<*(]G9&~uo^sȭDp+)JkT*"C>% %d˨ڢBkzxJD,/(;xR=F| Q #6Qv7G3MP_[jINh̅pW`Әg9^i_<y>Hh`v5s8uwkoLOÏW}y ShGDPb >~[7[iy/Uyx)rnY:XO;Pb\[34$oow߼vW}Bq Js -qrD p焷=گ{U@xxgǁ+$ ybc@zeBȐma!sV[ *o}@Q|9蓉᳧r~FCRDEM#W J8\@ wHUdDy6˹mj#4Ca^UC,nͽt>`/" ld3\sEVY 󥜂zV zw/|R9OՕIEl}7:xl=F4ӷK(oUyê|埃V 7vv _3xdUkV/h||tww'$WRZ]!h.р[n^t$6}lbogxIn&q~ym6UE{ffj]}7?t'n 5y}6iusq3-5y݁Ɛ$|[h?tHl+x07i5wݾܓu;0B1bz{݉rk (諿(rӋ = .zXiGf]Mx䝠rP+*ijʱR{ av]v'yػɣ~{:B;Ax_*57eF1I˸^ J'S͍qo? |^]2M??εqN T/Ay0%ѧI+vthڭ >2e/AK&wݞ3ɓi{D6nmH_\Ddy)*tӼ5od{yI*ißWQ[&C;X+W;'4 W$"IhuHc%Ƀ3y›0FKG+b V_K:e'I @C:Ť1BW6Nd WF&_m*s[EV WQW5-:# ;x RkO˘+d%2E9X?e99>:"keѼ_a`O:8^ց2(fU:rp}c- !qdXRD2 fDD~"K5 Sq ɚ5^`="4(EH -O&8t>lo89=g &"շ֫oVr#t5w9h>7%*$F# JIK<'1z]Yo9+_YxEan6yxڲVG{1}UKR2KdL~d0I~_r~;۷7N&Y7%vc|Mɓ~sNk9d(jegH.*SW\lHf)@*0 oI(EQDdPq6!"uZZA?ԊNJD6{Ex䜫j.:*Ěb>[.AE-L.%I^Yv98ǂ9eζUnǰ:vvBSqbɔK_q#fQ;i!(5̥󥺢Enz,2m[1kGn9 S-LEH=/M)NVJHW%Add`/_f,drf!k8^`/*8D㴞E:Ф-9/]e-~gfds8- ݝdU{){y)7W>ZVى5c~~[hC8Vс8^hw2<-?MDEA> QJ5Y;7D]yg`}s>ODQʍ$7 %J"{]z4uB{OOLD$gVIAɥ\S,mԛ(s8h*V.9RVqFQ|(!bD}*!kPf= Gp 9sjPȾ?/mDN6NiN!2k!CϴA-X8)N"{;1g{> ڱQo5s9s 1u@pP U>D7 Ĉ)OS\XŮ_VNs"JZj5f{AGMNG}b1^z0aݥ,dz~7뺥`i?,oY$<=owEBց~jB.I?9g9欦AnJ',>8ؾ2Q'$B;͢xXꯞ#ͤ5VӃXxZDD]`&듳EŞ)وx~3&_3"ꅅbDPDӠFKe{BGgImb y-%ş=بw*G-0[Xx탢Ƚ2GKwl<'"}[y 3ݟ _EKdIgs6Ӻ{$`qT.V߿ߛgt![+,|!%2BnoC(xHz0~e,$ί& X|>7S 嗛/5ݶçCǂ{|F&@/ɋo鄾]<}kZQlHGhI¡=ضIyu^:1M뻎W;ɝ?ޕhڦK{%$%F%Kx~0.9'~}:eZ^㿺1`x1K'tiɟ\!MCsB#s\r mir nѧpFrʯhub0r_my\Y:PvvS h컒yP ^K{~Ep'&|w?~Yύ.5X;ZzD kɞԿ| >YQ)!p>g j"v7sf4\@%(+q }5;}ndNu==H6riH(woi|:Xk۵#NI:ۯ_ZsFútJEBqE+ C6dU6DR@zMX{_<TI(«O7O#[)Mޅ[_vZ٢JgbHX+#3s\A)o &gצ.ySn#+%Ju:. M;>䏣T/7l$9 C.yh|;+mw#sz[KiSF{kkˇ/ l?EٮT[9osퟧW9P䤤'Sa9/3.!U&19A?φ _omKKFs:e'^jo,N' z'zvh5`kgȷ'f[w^SEvl{00j_%=l;)m- %o_(>0R)2:!HҵD㣓1'Wɓd14֨dt2`uAD1! )eh̬ӓ=aI}Ct{X[qs䜊LJE_@e#ZhjrtTGs,YUeHZNl ƤO2:\sGn)„#=}%Dbq.CHNtG& ׌MI%}1LZ"%! 5bf824fJN q1/cM֠MnyϋrEM%(VNl "?P뀔f4ZRѳAKFz|2#;ޞ;l$X/rec6Y  ' q{+ 2bJ"Z`S]{k/.5C;Qo鴗m[?Hul,3oNrm ycKK߻Rsu_k^ZfR74zoǭ8x|̣S&AOTD.Ku=tr=η`ʍn}Pґ}` PCA4N>Nӏz`*z`DaӓCșH=a9,yO#8gccY)Ҵ1ˢt<6hS*>~܅8Wڕx8en2"$PBĈvUB==-̼n,V|pwZ5@vuy_sYܶʐ\(): _ ]La9WQOI+ZƱZow{2`ejq9D%x(x Q +ѥ.?7S<9w]JJ>փ 7ւI%Rj}M':tOn{(6s~74J#ޛ#[K JgkL"5e_cT\+Wp B:C:qEghR\E;y JUvX66#*KP'-^El}ٕ`5\]N΋z}d*z&]"Q¶Lbjb#-EK>C{lRʲ^Fm=g9D숹fr=ƏUұZlUo[:-1XFry>Ǝ#(fcѼNy{\{ńߖܧ̚ʟV֖- ;Ojf 귿jUr,~Ӑ?p҈n|[ͫs89u2pkñ)(Xο߭10c(|k/,NΖVup&ёd5`O|lFj%R[wEFmzv?3u`m^q^a .2eOlO#ڙTRwo\HI*.ŵᎹh TRplɖQ*[|soW&bxw(q9D~x{s8tw7boEէO%SE^(j7gɠݜϦFMCfrtjZ2ChI˃U= 1"\ڜ̃39cK֟J}`QFHZFz}zvZjC>wpY^I3~wO>I(/師^.)wɊ|Iֆ!2SorP҆X*mjYyj#N0?\p=YRs;l!vvϫK ]Wz5_+imjf?gqcΎc? w\ 䌈 ~&XRo{O%%qT`ܯ`JF;KKE|qQ޳5"B X4^qcQ  ༀ|) Umz^ [ q+l^R~A$gBH6`@u#?Ĵe}Yԋk90 F:p?4YLH($zUh !E0A:*D]H;FQ$]Fn-^&(Zh!_Eﴊ ZIڗHtTy5>i6+t(4G@[.фiAPtzO5KzuK oVt<ϟNȽoUr&"Jh̵CZHu (}wi36Q}զZFf`XRII @&TKңF)2B: w˸(DUj49QQ"lق} ĊjA[ N(=.\ڃs$=M>j5VØS+%cyصm弁 1ȰߘS1#0ץϺZv17@c"Pݺ\0`!'P:x ؏ {^p#E'Kg+<`Z63ܖn#!{ӑh a3kۆmzE0#͋6s'^~QALr;o'Z0X85xsv%3;` |L;qQ?' d-r#26:Ej7~QgBDKײ%?껳&ɧP.Շhn\8[LZH&m9!3_+JNh E Pa@7Qz" xL!WQBJFJ\]Drq+mŪrEhuԡ\L XR5XJx gr.*KLus]Y1ڡC3y(;>CT_:$A.rS4=6^XtXLʝ5Xۃ[vcoTpiϯQz~u֔եGϯn.=Xa c~#~u'&r:Kb-N(E;%!&?"gWqz.pg$, J˔#_Ɨ B 0&7Q r$Lq7_/YrO DxD,<ꬎ;<:~(ªj7Jy-D<r^Ҋ~BQuJQg] 0PWT7fN(=>BaUG(w#*@ i2>HzW z1ZTS(BeT_px{2ȒJ(CiEFi\Mkv(]ҊjPzl( ovar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003414112115157464265017711 0ustar rootrootMar 21 08:27:44 crc systemd[1]: Starting Kubernetes Kubelet... Mar 21 08:27:44 crc restorecon[4695]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 08:27:44 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 08:27:45 crc restorecon[4695]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 21 08:27:45 crc restorecon[4695]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 21 08:27:46 crc kubenswrapper[4696]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 21 08:27:46 crc kubenswrapper[4696]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 21 08:27:46 crc kubenswrapper[4696]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 21 08:27:46 crc kubenswrapper[4696]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 21 08:27:46 crc kubenswrapper[4696]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 21 08:27:46 crc kubenswrapper[4696]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.322789 4696 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329017 4696 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329066 4696 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329075 4696 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329082 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329089 4696 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329094 4696 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329099 4696 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329104 4696 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329109 4696 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329114 4696 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329118 4696 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329126 4696 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329130 4696 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329136 4696 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329140 4696 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329146 4696 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329152 4696 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329158 4696 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329164 4696 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329170 4696 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329176 4696 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329184 4696 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329191 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329198 4696 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329205 4696 feature_gate.go:330] unrecognized feature gate: Example Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329211 4696 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329216 4696 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329221 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329226 4696 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329231 4696 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329236 4696 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329241 4696 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329246 4696 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329251 4696 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329256 4696 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329261 4696 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329266 4696 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329271 4696 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329276 4696 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329282 4696 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329287 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329292 4696 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329296 4696 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329301 4696 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329307 4696 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329313 4696 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329319 4696 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329325 4696 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329332 4696 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329338 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329343 4696 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329348 4696 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329353 4696 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329358 4696 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329363 4696 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329368 4696 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329372 4696 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329377 4696 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329382 4696 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329386 4696 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329391 4696 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329397 4696 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329403 4696 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329409 4696 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329414 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329420 4696 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329424 4696 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329432 4696 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329438 4696 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329444 4696 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.329449 4696 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329547 4696 flags.go:64] FLAG: --address="0.0.0.0" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329558 4696 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329566 4696 flags.go:64] FLAG: --anonymous-auth="true" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329574 4696 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329581 4696 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329587 4696 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329595 4696 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329602 4696 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329608 4696 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329614 4696 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329620 4696 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329627 4696 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329634 4696 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329640 4696 flags.go:64] FLAG: --cgroup-root="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329645 4696 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329651 4696 flags.go:64] FLAG: --client-ca-file="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329656 4696 flags.go:64] FLAG: --cloud-config="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329661 4696 flags.go:64] FLAG: --cloud-provider="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329667 4696 flags.go:64] FLAG: --cluster-dns="[]" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329674 4696 flags.go:64] FLAG: --cluster-domain="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329680 4696 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329685 4696 flags.go:64] FLAG: --config-dir="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329691 4696 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329697 4696 flags.go:64] FLAG: --container-log-max-files="5" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329704 4696 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329710 4696 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329715 4696 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329721 4696 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329727 4696 flags.go:64] FLAG: --contention-profiling="false" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329732 4696 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329738 4696 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329744 4696 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329749 4696 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329756 4696 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329762 4696 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329767 4696 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329773 4696 flags.go:64] FLAG: --enable-load-reader="false" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329778 4696 flags.go:64] FLAG: --enable-server="true" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329784 4696 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329790 4696 flags.go:64] FLAG: --event-burst="100" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329796 4696 flags.go:64] FLAG: --event-qps="50" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329803 4696 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329808 4696 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329814 4696 flags.go:64] FLAG: --eviction-hard="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329839 4696 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329845 4696 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329851 4696 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329858 4696 flags.go:64] FLAG: --eviction-soft="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329864 4696 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329869 4696 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329875 4696 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329880 4696 flags.go:64] FLAG: --experimental-mounter-path="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329886 4696 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329891 4696 flags.go:64] FLAG: --fail-swap-on="true" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329897 4696 flags.go:64] FLAG: --feature-gates="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329903 4696 flags.go:64] FLAG: --file-check-frequency="20s" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329909 4696 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329915 4696 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329921 4696 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329926 4696 flags.go:64] FLAG: --healthz-port="10248" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329932 4696 flags.go:64] FLAG: --help="false" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329937 4696 flags.go:64] FLAG: --hostname-override="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329943 4696 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329950 4696 flags.go:64] FLAG: --http-check-frequency="20s" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329955 4696 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329961 4696 flags.go:64] FLAG: --image-credential-provider-config="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329966 4696 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329972 4696 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329978 4696 flags.go:64] FLAG: --image-service-endpoint="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329983 4696 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329989 4696 flags.go:64] FLAG: --kube-api-burst="100" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.329994 4696 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330000 4696 flags.go:64] FLAG: --kube-api-qps="50" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330006 4696 flags.go:64] FLAG: --kube-reserved="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330012 4696 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330017 4696 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330023 4696 flags.go:64] FLAG: --kubelet-cgroups="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330029 4696 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330035 4696 flags.go:64] FLAG: --lock-file="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330040 4696 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330045 4696 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330051 4696 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330059 4696 flags.go:64] FLAG: --log-json-split-stream="false" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330065 4696 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330071 4696 flags.go:64] FLAG: --log-text-split-stream="false" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330076 4696 flags.go:64] FLAG: --logging-format="text" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330081 4696 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330087 4696 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330093 4696 flags.go:64] FLAG: --manifest-url="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330098 4696 flags.go:64] FLAG: --manifest-url-header="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330105 4696 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330111 4696 flags.go:64] FLAG: --max-open-files="1000000" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330117 4696 flags.go:64] FLAG: --max-pods="110" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330124 4696 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330129 4696 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330134 4696 flags.go:64] FLAG: --memory-manager-policy="None" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330140 4696 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330146 4696 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330153 4696 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330159 4696 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330175 4696 flags.go:64] FLAG: --node-status-max-images="50" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330181 4696 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330187 4696 flags.go:64] FLAG: --oom-score-adj="-999" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330192 4696 flags.go:64] FLAG: --pod-cidr="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330198 4696 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330207 4696 flags.go:64] FLAG: --pod-manifest-path="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330212 4696 flags.go:64] FLAG: --pod-max-pids="-1" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330219 4696 flags.go:64] FLAG: --pods-per-core="0" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330225 4696 flags.go:64] FLAG: --port="10250" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330232 4696 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330238 4696 flags.go:64] FLAG: --provider-id="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330243 4696 flags.go:64] FLAG: --qos-reserved="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330249 4696 flags.go:64] FLAG: --read-only-port="10255" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330255 4696 flags.go:64] FLAG: --register-node="true" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330261 4696 flags.go:64] FLAG: --register-schedulable="true" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330266 4696 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330275 4696 flags.go:64] FLAG: --registry-burst="10" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330281 4696 flags.go:64] FLAG: --registry-qps="5" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330286 4696 flags.go:64] FLAG: --reserved-cpus="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330292 4696 flags.go:64] FLAG: --reserved-memory="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330299 4696 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330305 4696 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330310 4696 flags.go:64] FLAG: --rotate-certificates="false" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330316 4696 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330321 4696 flags.go:64] FLAG: --runonce="false" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330327 4696 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330333 4696 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330339 4696 flags.go:64] FLAG: --seccomp-default="false" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330345 4696 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330350 4696 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330356 4696 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330362 4696 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330368 4696 flags.go:64] FLAG: --storage-driver-password="root" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330374 4696 flags.go:64] FLAG: --storage-driver-secure="false" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330379 4696 flags.go:64] FLAG: --storage-driver-table="stats" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330385 4696 flags.go:64] FLAG: --storage-driver-user="root" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330390 4696 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330396 4696 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330402 4696 flags.go:64] FLAG: --system-cgroups="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330407 4696 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330416 4696 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330422 4696 flags.go:64] FLAG: --tls-cert-file="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330428 4696 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330435 4696 flags.go:64] FLAG: --tls-min-version="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330441 4696 flags.go:64] FLAG: --tls-private-key-file="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330446 4696 flags.go:64] FLAG: --topology-manager-policy="none" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330452 4696 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330458 4696 flags.go:64] FLAG: --topology-manager-scope="container" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330463 4696 flags.go:64] FLAG: --v="2" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330470 4696 flags.go:64] FLAG: --version="false" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330477 4696 flags.go:64] FLAG: --vmodule="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330483 4696 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.330489 4696 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330617 4696 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330624 4696 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330631 4696 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330636 4696 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330641 4696 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330645 4696 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330651 4696 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330656 4696 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330662 4696 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330668 4696 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330673 4696 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330679 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330684 4696 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330690 4696 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330696 4696 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330701 4696 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330706 4696 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330711 4696 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330718 4696 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330724 4696 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330729 4696 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330733 4696 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330738 4696 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330743 4696 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330748 4696 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330753 4696 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330758 4696 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330763 4696 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330767 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330772 4696 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330777 4696 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330781 4696 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330786 4696 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330791 4696 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330796 4696 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330800 4696 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330805 4696 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330811 4696 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330832 4696 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330837 4696 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330842 4696 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330847 4696 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330854 4696 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330860 4696 feature_gate.go:330] unrecognized feature gate: Example Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330867 4696 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330872 4696 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330877 4696 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330882 4696 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330887 4696 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330894 4696 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330900 4696 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330905 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330912 4696 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330918 4696 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330924 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330929 4696 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330934 4696 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330939 4696 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330944 4696 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330949 4696 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330954 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330959 4696 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330963 4696 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330968 4696 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330974 4696 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330980 4696 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330987 4696 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330991 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.330996 4696 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.331001 4696 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.331006 4696 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.331022 4696 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.347124 4696 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.347180 4696 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347344 4696 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347360 4696 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347370 4696 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347380 4696 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347389 4696 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347397 4696 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347405 4696 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347413 4696 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347421 4696 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347429 4696 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347440 4696 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347452 4696 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347461 4696 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347470 4696 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347479 4696 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347488 4696 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347496 4696 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347504 4696 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347512 4696 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347521 4696 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347531 4696 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347544 4696 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347554 4696 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347565 4696 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347576 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347587 4696 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347597 4696 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347608 4696 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347620 4696 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347630 4696 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347639 4696 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347647 4696 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347655 4696 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347666 4696 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347677 4696 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347687 4696 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347697 4696 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347708 4696 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347718 4696 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347728 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347738 4696 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347751 4696 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347765 4696 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347781 4696 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347793 4696 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347803 4696 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347853 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347865 4696 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347875 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347886 4696 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347896 4696 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347907 4696 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347918 4696 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347929 4696 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347940 4696 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347950 4696 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347960 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347970 4696 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347980 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347990 4696 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.347999 4696 feature_gate.go:330] unrecognized feature gate: Example Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348008 4696 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348021 4696 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348031 4696 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348041 4696 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348051 4696 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348060 4696 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348070 4696 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348081 4696 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348091 4696 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348103 4696 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.348121 4696 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348424 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348442 4696 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348451 4696 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348460 4696 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348470 4696 feature_gate.go:330] unrecognized feature gate: Example Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348479 4696 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348487 4696 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348495 4696 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348503 4696 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348514 4696 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348521 4696 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348529 4696 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348536 4696 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348545 4696 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348553 4696 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348560 4696 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348568 4696 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348576 4696 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348586 4696 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348597 4696 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348606 4696 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348616 4696 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348626 4696 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348639 4696 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348652 4696 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348663 4696 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348675 4696 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348685 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348695 4696 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348706 4696 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348716 4696 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348726 4696 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348735 4696 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348745 4696 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348755 4696 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348766 4696 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348776 4696 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348785 4696 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348798 4696 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348843 4696 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348855 4696 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348866 4696 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348874 4696 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348884 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348893 4696 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348905 4696 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348913 4696 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348921 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348929 4696 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348937 4696 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348945 4696 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348953 4696 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348961 4696 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348968 4696 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348976 4696 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348984 4696 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348991 4696 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.348999 4696 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.349009 4696 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.349019 4696 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.349028 4696 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.349064 4696 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.349074 4696 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.349082 4696 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.349164 4696 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.349174 4696 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.349182 4696 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.349191 4696 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.349200 4696 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.349208 4696 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.349217 4696 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.349230 4696 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.349527 4696 server.go:940] "Client rotation is on, will bootstrap in background" Mar 21 08:27:46 crc kubenswrapper[4696]: E0321 08:27:46.355099 4696 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.359033 4696 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.359163 4696 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.361020 4696 server.go:997] "Starting client certificate rotation" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.361114 4696 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.361318 4696 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.382104 4696 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.384375 4696 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 21 08:27:46 crc kubenswrapper[4696]: E0321 08:27:46.384507 4696 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.399424 4696 log.go:25] "Validated CRI v1 runtime API" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.429461 4696 log.go:25] "Validated CRI v1 image API" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.433707 4696 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.437741 4696 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-21-08-22-31-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.437783 4696 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:41 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:42 fsType:tmpfs blockSize:0}] Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.456627 4696 manager.go:217] Machine: {Timestamp:2026-03-21 08:27:46.454889634 +0000 UTC m=+0.575770377 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:fa0c4141-962b-4589-98ba-ecb07b216fa9 BootID:7ca94532-e7e5-44cd-8187-860d6872e1ea Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:41 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:42 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:68:0f:58 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:68:0f:58 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:35:b2:af Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:d2:10:ba Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:ab:c4:f2 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:14:80:95 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:46:6e:85:73:20:70 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:a2:5f:09:5e:99:42 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.456963 4696 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.457161 4696 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.457854 4696 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.458137 4696 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.458177 4696 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.458372 4696 topology_manager.go:138] "Creating topology manager with none policy" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.458383 4696 container_manager_linux.go:303] "Creating device plugin manager" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.458866 4696 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.458896 4696 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.459082 4696 state_mem.go:36] "Initialized new in-memory state store" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.459173 4696 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.463890 4696 kubelet.go:418] "Attempting to sync node with API server" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.463912 4696 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.463935 4696 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.463948 4696 kubelet.go:324] "Adding apiserver pod source" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.463958 4696 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.467399 4696 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.468273 4696 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.470144 4696 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.470188 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 21 08:27:46 crc kubenswrapper[4696]: E0321 08:27:46.470305 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.470492 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 21 08:27:46 crc kubenswrapper[4696]: E0321 08:27:46.470648 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.471462 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.471492 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.471503 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.471514 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.471530 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.471541 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.471550 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.471566 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.471579 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.471600 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.471626 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.471637 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.472402 4696 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.472896 4696 server.go:1280] "Started kubelet" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.473145 4696 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.474787 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 21 08:27:46 crc systemd[1]: Started Kubernetes Kubelet. Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.476440 4696 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.478229 4696 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.478313 4696 server.go:460] "Adding debug handlers to kubelet server" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.478756 4696 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.479019 4696 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.479174 4696 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.479220 4696 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 21 08:27:46 crc kubenswrapper[4696]: E0321 08:27:46.479264 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.479371 4696 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.484167 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 21 08:27:46 crc kubenswrapper[4696]: E0321 08:27:46.484258 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 21 08:27:46 crc kubenswrapper[4696]: E0321 08:27:46.483918 4696 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.174:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189ecde04ca5f39c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.472866716 +0000 UTC m=+0.593747449,LastTimestamp:2026-03-21 08:27:46.472866716 +0000 UTC m=+0.593747449,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.485557 4696 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.485607 4696 factory.go:55] Registering systemd factory Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.485621 4696 factory.go:221] Registration of the systemd container factory successfully Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.486356 4696 factory.go:153] Registering CRI-O factory Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.486383 4696 factory.go:221] Registration of the crio container factory successfully Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.486429 4696 factory.go:103] Registering Raw factory Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.487170 4696 manager.go:1196] Started watching for new ooms in manager Mar 21 08:27:46 crc kubenswrapper[4696]: E0321 08:27:46.486493 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="200ms" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.490273 4696 manager.go:319] Starting recovery of all containers Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.496960 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497085 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497127 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497141 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497154 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497167 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497180 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497195 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497210 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497223 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497285 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497298 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497311 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497329 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497342 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497357 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497376 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497394 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497421 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497439 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497456 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497475 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497492 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497508 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497533 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497553 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497580 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497629 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497646 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497664 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497681 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497697 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497716 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497734 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497751 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497770 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497788 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497806 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497855 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497875 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497895 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497917 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497939 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497959 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.497977 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498024 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498043 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498060 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498077 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498094 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498115 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498132 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498155 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498175 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498194 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498213 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498234 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498256 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498274 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498293 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498311 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498331 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498349 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498370 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498388 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498406 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498423 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498441 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498458 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498474 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498492 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498510 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498527 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498546 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498564 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498581 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498600 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498619 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498637 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498655 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498672 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498690 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498708 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498726 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498744 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498762 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498780 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498799 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498846 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498869 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498888 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498906 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498922 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498935 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498949 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498974 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.498988 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.499002 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.499015 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.499029 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.499043 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.499062 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.499076 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.499089 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.499110 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.499129 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.499144 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.499159 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.499177 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.499196 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.500855 4696 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.500897 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.500921 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.500941 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.500963 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501010 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501030 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501050 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501068 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501087 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501104 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501126 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501160 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501188 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501208 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501227 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501245 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501273 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501293 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501312 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501332 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501351 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501369 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501414 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501434 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501452 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501469 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501490 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501509 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501527 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501546 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501566 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501584 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501597 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501611 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501623 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501636 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501652 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501665 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501678 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501692 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501705 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501718 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501730 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501742 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501754 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501766 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501778 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501791 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501808 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501974 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.501992 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502010 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502026 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502042 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502060 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502078 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502095 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502113 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502140 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502156 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502173 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502192 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502210 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502229 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502246 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502266 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502283 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502300 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502323 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502338 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502357 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502376 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502392 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502410 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502431 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502449 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502467 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502487 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502509 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502527 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502546 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502576 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502595 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502613 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502630 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502645 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502659 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502672 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502685 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502697 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502712 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502724 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502738 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502752 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502787 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502800 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502839 4696 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502857 4696 reconstruct.go:97] "Volume reconstruction finished" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.502869 4696 reconciler.go:26] "Reconciler: start to sync state" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.526329 4696 manager.go:324] Recovery completed Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.529756 4696 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.532664 4696 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.533329 4696 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.533377 4696 kubelet.go:2335] "Starting kubelet main sync loop" Mar 21 08:27:46 crc kubenswrapper[4696]: E0321 08:27:46.533484 4696 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 21 08:27:46 crc kubenswrapper[4696]: W0321 08:27:46.536586 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 21 08:27:46 crc kubenswrapper[4696]: E0321 08:27:46.536746 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.538173 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.539447 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.539469 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.539477 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.540053 4696 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.540066 4696 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.540112 4696 state_mem.go:36] "Initialized new in-memory state store" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.550210 4696 policy_none.go:49] "None policy: Start" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.550905 4696 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.550930 4696 state_mem.go:35] "Initializing new in-memory state store" Mar 21 08:27:46 crc kubenswrapper[4696]: E0321 08:27:46.579548 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.599358 4696 manager.go:334] "Starting Device Plugin manager" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.599459 4696 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.599484 4696 server.go:79] "Starting device plugin registration server" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.600267 4696 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.600302 4696 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.601063 4696 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.601216 4696 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.601238 4696 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 21 08:27:46 crc kubenswrapper[4696]: E0321 08:27:46.608683 4696 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.633896 4696 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.634034 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.635153 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.635183 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.635191 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.635311 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.635984 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.636002 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.636010 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.636261 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.636283 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.636325 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.636622 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.636649 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.637335 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.637349 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.637356 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.637440 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.637735 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.637759 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.638118 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.638143 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.638154 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.638161 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.638188 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.638205 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.638292 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.638304 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.638311 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.638412 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.638422 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.638438 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.638514 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.638880 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.638900 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.639915 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.639939 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.639954 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.639959 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.639976 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.639987 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.640082 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.640108 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.640735 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.640751 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.640758 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:46 crc kubenswrapper[4696]: E0321 08:27:46.688922 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="400ms" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.701310 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.702943 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.702990 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.703037 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.703076 4696 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 08:27:46 crc kubenswrapper[4696]: E0321 08:27:46.703890 4696 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.706702 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.706738 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.706768 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.706790 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.706834 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.706907 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.706972 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.707003 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.707050 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.707116 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.707167 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.707185 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.707210 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.707317 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.707398 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.808801 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.808959 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.808998 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809057 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809084 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809108 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809174 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809132 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809118 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809225 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809183 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809307 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809265 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809426 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809473 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809502 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809526 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809543 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809556 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809594 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809597 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809616 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809548 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809639 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809683 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809757 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809770 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809804 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809855 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.809878 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.904581 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.907211 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.907253 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.907271 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.907302 4696 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 08:27:46 crc kubenswrapper[4696]: E0321 08:27:46.907907 4696 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.971389 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 21 08:27:46 crc kubenswrapper[4696]: I0321 08:27:46.995586 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 08:27:47 crc kubenswrapper[4696]: I0321 08:27:47.006106 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:27:47 crc kubenswrapper[4696]: W0321 08:27:47.025192 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-4863ef5a9cec1122d3b0c3b4a8efa808c359e5a448e6744c97ed8f0dfd06895e WatchSource:0}: Error finding container 4863ef5a9cec1122d3b0c3b4a8efa808c359e5a448e6744c97ed8f0dfd06895e: Status 404 returned error can't find the container with id 4863ef5a9cec1122d3b0c3b4a8efa808c359e5a448e6744c97ed8f0dfd06895e Mar 21 08:27:47 crc kubenswrapper[4696]: I0321 08:27:47.028025 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:27:47 crc kubenswrapper[4696]: I0321 08:27:47.031712 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 08:27:47 crc kubenswrapper[4696]: W0321 08:27:47.032270 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-ee741f0c2aaeb0379655432b29ac6643a7359c19ec71bc9750482c267b756b04 WatchSource:0}: Error finding container ee741f0c2aaeb0379655432b29ac6643a7359c19ec71bc9750482c267b756b04: Status 404 returned error can't find the container with id ee741f0c2aaeb0379655432b29ac6643a7359c19ec71bc9750482c267b756b04 Mar 21 08:27:47 crc kubenswrapper[4696]: W0321 08:27:47.034999 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-cf2b4e2880720248760752e3d40108b0f4c5d8aaa45c6b7c83df1c8a89ee82e8 WatchSource:0}: Error finding container cf2b4e2880720248760752e3d40108b0f4c5d8aaa45c6b7c83df1c8a89ee82e8: Status 404 returned error can't find the container with id cf2b4e2880720248760752e3d40108b0f4c5d8aaa45c6b7c83df1c8a89ee82e8 Mar 21 08:27:47 crc kubenswrapper[4696]: W0321 08:27:47.057111 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-8a2ab34fad7ca38ec45da5654f800c2abc40993365831ca26e3d4c595ee7cf33 WatchSource:0}: Error finding container 8a2ab34fad7ca38ec45da5654f800c2abc40993365831ca26e3d4c595ee7cf33: Status 404 returned error can't find the container with id 8a2ab34fad7ca38ec45da5654f800c2abc40993365831ca26e3d4c595ee7cf33 Mar 21 08:27:47 crc kubenswrapper[4696]: W0321 08:27:47.058993 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-d47d10a511ba240d7b9f299092f02661f01d902e1517fee769f711112f687dc4 WatchSource:0}: Error finding container d47d10a511ba240d7b9f299092f02661f01d902e1517fee769f711112f687dc4: Status 404 returned error can't find the container with id d47d10a511ba240d7b9f299092f02661f01d902e1517fee769f711112f687dc4 Mar 21 08:27:47 crc kubenswrapper[4696]: E0321 08:27:47.090741 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="800ms" Mar 21 08:27:47 crc kubenswrapper[4696]: I0321 08:27:47.308699 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:47 crc kubenswrapper[4696]: I0321 08:27:47.310843 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:47 crc kubenswrapper[4696]: I0321 08:27:47.310880 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:47 crc kubenswrapper[4696]: I0321 08:27:47.310891 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:47 crc kubenswrapper[4696]: I0321 08:27:47.310917 4696 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 08:27:47 crc kubenswrapper[4696]: E0321 08:27:47.311435 4696 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Mar 21 08:27:47 crc kubenswrapper[4696]: W0321 08:27:47.375079 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 21 08:27:47 crc kubenswrapper[4696]: E0321 08:27:47.375161 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 21 08:27:47 crc kubenswrapper[4696]: W0321 08:27:47.377990 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 21 08:27:47 crc kubenswrapper[4696]: E0321 08:27:47.378046 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 21 08:27:47 crc kubenswrapper[4696]: I0321 08:27:47.476247 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 21 08:27:47 crc kubenswrapper[4696]: I0321 08:27:47.537491 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8a2ab34fad7ca38ec45da5654f800c2abc40993365831ca26e3d4c595ee7cf33"} Mar 21 08:27:47 crc kubenswrapper[4696]: I0321 08:27:47.538662 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d47d10a511ba240d7b9f299092f02661f01d902e1517fee769f711112f687dc4"} Mar 21 08:27:47 crc kubenswrapper[4696]: I0321 08:27:47.539478 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cf2b4e2880720248760752e3d40108b0f4c5d8aaa45c6b7c83df1c8a89ee82e8"} Mar 21 08:27:47 crc kubenswrapper[4696]: I0321 08:27:47.540416 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ee741f0c2aaeb0379655432b29ac6643a7359c19ec71bc9750482c267b756b04"} Mar 21 08:27:47 crc kubenswrapper[4696]: I0321 08:27:47.541221 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4863ef5a9cec1122d3b0c3b4a8efa808c359e5a448e6744c97ed8f0dfd06895e"} Mar 21 08:27:47 crc kubenswrapper[4696]: E0321 08:27:47.892155 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="1.6s" Mar 21 08:27:47 crc kubenswrapper[4696]: W0321 08:27:47.900724 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 21 08:27:47 crc kubenswrapper[4696]: E0321 08:27:47.900929 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 21 08:27:47 crc kubenswrapper[4696]: W0321 08:27:47.911663 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 21 08:27:47 crc kubenswrapper[4696]: E0321 08:27:47.911869 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.112511 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.113804 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.113873 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.113887 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.113913 4696 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 08:27:48 crc kubenswrapper[4696]: E0321 08:27:48.114409 4696 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.476458 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.534487 4696 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 21 08:27:48 crc kubenswrapper[4696]: E0321 08:27:48.535924 4696 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.546622 4696 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="9feeb272bdbb29ae95036191612aedd3291adb6f0048451607c0d3d7cd5bc4ca" exitCode=0 Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.546695 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"9feeb272bdbb29ae95036191612aedd3291adb6f0048451607c0d3d7cd5bc4ca"} Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.546995 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.549102 4696 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fede52f3fbd397eb207a4d0daa4587e1987bcd31c175d212557666a4e939701f" exitCode=0 Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.549175 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"fede52f3fbd397eb207a4d0daa4587e1987bcd31c175d212557666a4e939701f"} Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.549128 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.549245 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.549266 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.549245 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.550255 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.550291 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.550302 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.551791 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.552673 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.552698 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.552709 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.553008 4696 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="a8d6120e42d2d5ec71cb0defd7b303ff9faa1888f6b993465177b4df178f4ce4" exitCode=0 Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.553127 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"a8d6120e42d2d5ec71cb0defd7b303ff9faa1888f6b993465177b4df178f4ce4"} Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.553204 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.554019 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.554058 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.554071 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.557344 4696 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="92eb64987eafae1fe0d928d223b375fc5bf64e573cc161026ab1c35d66be5cec" exitCode=0 Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.557433 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"92eb64987eafae1fe0d928d223b375fc5bf64e573cc161026ab1c35d66be5cec"} Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.557552 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.562173 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.562214 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.562227 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.566694 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9b4e59dfd2f354e290ba0fa628127aab9fc3d2d26009bfbbd2f2e5c6d72ea687"} Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.566777 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a68de4a2c70abcf12ca6cfb362768774f880fb74456aacb1585fb7d6ac29466d"} Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.566795 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.566799 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6f9f2c48d413fd10728233381f94e3fb94d5b73000d1337244a76a895b2bcb31"} Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.567466 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f4111d26588a9cb6e28559a10dcc9a940428b646a3072500c455e099ab8eb5ee"} Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.569077 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.569104 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:48 crc kubenswrapper[4696]: I0321 08:27:48.569129 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:49 crc kubenswrapper[4696]: W0321 08:27:49.087196 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 21 08:27:49 crc kubenswrapper[4696]: E0321 08:27:49.087357 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.476977 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 21 08:27:49 crc kubenswrapper[4696]: E0321 08:27:49.496905 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="3.2s" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.571247 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.571289 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"3326095ddcdc1a4c02ddbb1e1bd084a8eadf6f75a89289f6a7c0038aeaecdd6e"} Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.572464 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.572504 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.572516 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.573670 4696 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="0ee49ba00e2edcd6cff8e5d17009526a9c71a3007ab31d1e2e704be8d321cf06" exitCode=0 Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.573779 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"0ee49ba00e2edcd6cff8e5d17009526a9c71a3007ab31d1e2e704be8d321cf06"} Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.573800 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.574450 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.574471 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.574479 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.578399 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a3489a261840efbd8344b1560b3430898fb0ce389fcea5c1578481628ed58863"} Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.578428 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d3be3dcf8330e7555e067dcfd531444e9d4aa84dbc7e27936647e15cd25b4d31"} Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.578438 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"168622b38b3c0050827b1fc8cc906e67d90ea98512a67b676e0828095c3bc24f"} Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.578531 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.580145 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.580208 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.580227 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.588551 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"824fb0d35d4bdbe666975fd095c6299d742b56dc6c34b2ac87cffd9708c821aa"} Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.588614 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ac98b604eeaf478083bc6291ecded1c1c68b1bb7a844b281c66447fb9b7ff248"} Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.588634 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e83144c8a489a404f9916385b8216651b6fcfa968e843b43b62901adbaab80f1"} Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.588637 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.588649 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f72b31e3f8525ae3cbfa7a2b4ccbb88d4aac21c4e1e802d2cde33aa4408e5b24"} Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.589756 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.589873 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.589888 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.715254 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.716538 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.716629 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.716645 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.716672 4696 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 08:27:49 crc kubenswrapper[4696]: E0321 08:27:49.717286 4696 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Mar 21 08:27:49 crc kubenswrapper[4696]: I0321 08:27:49.983460 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 08:27:50 crc kubenswrapper[4696]: I0321 08:27:50.597637 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"573be38366713e9e429e2d24bef5f6c0f923de7715311fc10dfcae6a0e4a4c00"} Mar 21 08:27:50 crc kubenswrapper[4696]: I0321 08:27:50.597802 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:50 crc kubenswrapper[4696]: I0321 08:27:50.598979 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:50 crc kubenswrapper[4696]: I0321 08:27:50.599027 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:50 crc kubenswrapper[4696]: I0321 08:27:50.599043 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:50 crc kubenswrapper[4696]: I0321 08:27:50.605734 4696 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9673c732c1260c675a372565737e693e7449576f4615457e2e2f5c3099048bbd" exitCode=0 Mar 21 08:27:50 crc kubenswrapper[4696]: I0321 08:27:50.605873 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:50 crc kubenswrapper[4696]: I0321 08:27:50.605869 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9673c732c1260c675a372565737e693e7449576f4615457e2e2f5c3099048bbd"} Mar 21 08:27:50 crc kubenswrapper[4696]: I0321 08:27:50.606003 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:50 crc kubenswrapper[4696]: I0321 08:27:50.606512 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:50 crc kubenswrapper[4696]: I0321 08:27:50.606915 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:50 crc kubenswrapper[4696]: I0321 08:27:50.606973 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:50 crc kubenswrapper[4696]: I0321 08:27:50.606997 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:50 crc kubenswrapper[4696]: I0321 08:27:50.606980 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:50 crc kubenswrapper[4696]: I0321 08:27:50.607056 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:50 crc kubenswrapper[4696]: I0321 08:27:50.607011 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:50 crc kubenswrapper[4696]: I0321 08:27:50.608003 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:50 crc kubenswrapper[4696]: I0321 08:27:50.608028 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:50 crc kubenswrapper[4696]: I0321 08:27:50.608036 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:51 crc kubenswrapper[4696]: I0321 08:27:51.215111 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:27:51 crc kubenswrapper[4696]: I0321 08:27:51.614786 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fdc2f3a23a214f859a464a816287f7a390b42caab779e9731d8487859eb754c8"} Mar 21 08:27:51 crc kubenswrapper[4696]: I0321 08:27:51.614922 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:51 crc kubenswrapper[4696]: I0321 08:27:51.614918 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:51 crc kubenswrapper[4696]: I0321 08:27:51.615000 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e6adeaa8bf6c16eb2b49fcbef5f6876bac0ba3cb134ac4e85aeae86703cc65a3"} Mar 21 08:27:51 crc kubenswrapper[4696]: I0321 08:27:51.615127 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"48e210710728ef3fbd773cb3d7dc98251e389c18e1858e1f9aaee6edf736ea9e"} Mar 21 08:27:51 crc kubenswrapper[4696]: I0321 08:27:51.615156 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ab9c28c6575ba7afc2759d51e29d1288b42a9f91d8b092c77b77790b2f72baa5"} Mar 21 08:27:51 crc kubenswrapper[4696]: I0321 08:27:51.616220 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:51 crc kubenswrapper[4696]: I0321 08:27:51.616263 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:51 crc kubenswrapper[4696]: I0321 08:27:51.616280 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:51 crc kubenswrapper[4696]: I0321 08:27:51.616471 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:51 crc kubenswrapper[4696]: I0321 08:27:51.616509 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:51 crc kubenswrapper[4696]: I0321 08:27:51.616520 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.100414 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.100619 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.102039 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.102079 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.102098 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.148962 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.623728 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.624032 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4ab219afc75f24951063b1288aea0008adc049e32d821c0f174479dee5ccdde1"} Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.624184 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.624187 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.626256 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.626302 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.626316 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.627131 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.627171 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.627183 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.627477 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.627638 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.627779 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.918398 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.920783 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.920890 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.920916 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.920965 4696 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 08:27:52 crc kubenswrapper[4696]: I0321 08:27:52.929959 4696 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 21 08:27:53 crc kubenswrapper[4696]: I0321 08:27:53.626616 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:53 crc kubenswrapper[4696]: I0321 08:27:53.627768 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:53 crc kubenswrapper[4696]: I0321 08:27:53.627862 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:53 crc kubenswrapper[4696]: I0321 08:27:53.627881 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:53 crc kubenswrapper[4696]: I0321 08:27:53.769541 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:27:53 crc kubenswrapper[4696]: I0321 08:27:53.769748 4696 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 21 08:27:53 crc kubenswrapper[4696]: I0321 08:27:53.769801 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:53 crc kubenswrapper[4696]: I0321 08:27:53.771539 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:53 crc kubenswrapper[4696]: I0321 08:27:53.771662 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:53 crc kubenswrapper[4696]: I0321 08:27:53.771686 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:54 crc kubenswrapper[4696]: I0321 08:27:54.231879 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:27:54 crc kubenswrapper[4696]: I0321 08:27:54.232108 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:54 crc kubenswrapper[4696]: I0321 08:27:54.233643 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:54 crc kubenswrapper[4696]: I0321 08:27:54.233724 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:54 crc kubenswrapper[4696]: I0321 08:27:54.233748 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:54 crc kubenswrapper[4696]: I0321 08:27:54.890906 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:27:54 crc kubenswrapper[4696]: I0321 08:27:54.891150 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:54 crc kubenswrapper[4696]: I0321 08:27:54.892680 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:54 crc kubenswrapper[4696]: I0321 08:27:54.892715 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:54 crc kubenswrapper[4696]: I0321 08:27:54.892726 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:55 crc kubenswrapper[4696]: I0321 08:27:55.133262 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:27:55 crc kubenswrapper[4696]: I0321 08:27:55.133563 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:55 crc kubenswrapper[4696]: I0321 08:27:55.135240 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:55 crc kubenswrapper[4696]: I0321 08:27:55.135304 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:55 crc kubenswrapper[4696]: I0321 08:27:55.135322 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:56 crc kubenswrapper[4696]: I0321 08:27:56.036800 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:27:56 crc kubenswrapper[4696]: I0321 08:27:56.037108 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:56 crc kubenswrapper[4696]: I0321 08:27:56.038543 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:56 crc kubenswrapper[4696]: I0321 08:27:56.038592 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:56 crc kubenswrapper[4696]: I0321 08:27:56.038612 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:56 crc kubenswrapper[4696]: E0321 08:27:56.609059 4696 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 08:27:57 crc kubenswrapper[4696]: I0321 08:27:57.051415 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 21 08:27:57 crc kubenswrapper[4696]: I0321 08:27:57.051661 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:27:57 crc kubenswrapper[4696]: I0321 08:27:57.053258 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:27:57 crc kubenswrapper[4696]: I0321 08:27:57.053308 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:27:57 crc kubenswrapper[4696]: I0321 08:27:57.053326 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:27:59 crc kubenswrapper[4696]: I0321 08:27:59.037903 4696 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 21 08:27:59 crc kubenswrapper[4696]: I0321 08:27:59.037990 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 21 08:28:00 crc kubenswrapper[4696]: W0321 08:28:00.087513 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 21 08:28:00 crc kubenswrapper[4696]: I0321 08:28:00.087634 4696 trace.go:236] Trace[1211929602]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Mar-2026 08:27:50.086) (total time: 10001ms): Mar 21 08:28:00 crc kubenswrapper[4696]: Trace[1211929602]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (08:28:00.087) Mar 21 08:28:00 crc kubenswrapper[4696]: Trace[1211929602]: [10.001420918s] [10.001420918s] END Mar 21 08:28:00 crc kubenswrapper[4696]: E0321 08:28:00.087665 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 21 08:28:00 crc kubenswrapper[4696]: I0321 08:28:00.375777 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:00Z is after 2026-02-23T05:33:13Z Mar 21 08:28:00 crc kubenswrapper[4696]: W0321 08:28:00.376197 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:00Z is after 2026-02-23T05:33:13Z Mar 21 08:28:00 crc kubenswrapper[4696]: E0321 08:28:00.376300 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:00Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 21 08:28:00 crc kubenswrapper[4696]: W0321 08:28:00.379658 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:00Z is after 2026-02-23T05:33:13Z Mar 21 08:28:00 crc kubenswrapper[4696]: E0321 08:28:00.379803 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:00Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 21 08:28:00 crc kubenswrapper[4696]: W0321 08:28:00.383203 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:00Z is after 2026-02-23T05:33:13Z Mar 21 08:28:00 crc kubenswrapper[4696]: E0321 08:28:00.383315 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:00Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 21 08:28:00 crc kubenswrapper[4696]: E0321 08:28:00.385597 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:00Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 21 08:28:00 crc kubenswrapper[4696]: E0321 08:28:00.387391 4696 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:00Z is after 2026-02-23T05:33:13Z" node="crc" Mar 21 08:28:00 crc kubenswrapper[4696]: E0321 08:28:00.387900 4696 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:00Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 21 08:28:00 crc kubenswrapper[4696]: E0321 08:28:00.390741 4696 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:00Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189ecde04ca5f39c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.472866716 +0000 UTC m=+0.593747449,LastTimestamp:2026-03-21 08:27:46.472866716 +0000 UTC m=+0.593747449,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:00 crc kubenswrapper[4696]: I0321 08:28:00.397707 4696 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 21 08:28:00 crc kubenswrapper[4696]: I0321 08:28:00.397801 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 21 08:28:00 crc kubenswrapper[4696]: I0321 08:28:00.404235 4696 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 21 08:28:00 crc kubenswrapper[4696]: I0321 08:28:00.404330 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 21 08:28:00 crc kubenswrapper[4696]: I0321 08:28:00.478487 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:00Z is after 2026-02-23T05:33:13Z Mar 21 08:28:00 crc kubenswrapper[4696]: I0321 08:28:00.646521 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 21 08:28:00 crc kubenswrapper[4696]: I0321 08:28:00.648225 4696 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="573be38366713e9e429e2d24bef5f6c0f923de7715311fc10dfcae6a0e4a4c00" exitCode=255 Mar 21 08:28:00 crc kubenswrapper[4696]: I0321 08:28:00.648283 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"573be38366713e9e429e2d24bef5f6c0f923de7715311fc10dfcae6a0e4a4c00"} Mar 21 08:28:00 crc kubenswrapper[4696]: I0321 08:28:00.648508 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:00 crc kubenswrapper[4696]: I0321 08:28:00.649471 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:00 crc kubenswrapper[4696]: I0321 08:28:00.649526 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:00 crc kubenswrapper[4696]: I0321 08:28:00.649542 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:00 crc kubenswrapper[4696]: I0321 08:28:00.650381 4696 scope.go:117] "RemoveContainer" containerID="573be38366713e9e429e2d24bef5f6c0f923de7715311fc10dfcae6a0e4a4c00" Mar 21 08:28:01 crc kubenswrapper[4696]: I0321 08:28:01.245017 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 21 08:28:01 crc kubenswrapper[4696]: I0321 08:28:01.245221 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:01 crc kubenswrapper[4696]: I0321 08:28:01.246639 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:01 crc kubenswrapper[4696]: I0321 08:28:01.246671 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:01 crc kubenswrapper[4696]: I0321 08:28:01.246680 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:01 crc kubenswrapper[4696]: I0321 08:28:01.294411 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 21 08:28:01 crc kubenswrapper[4696]: I0321 08:28:01.491162 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:01Z is after 2026-02-23T05:33:13Z Mar 21 08:28:01 crc kubenswrapper[4696]: I0321 08:28:01.654183 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 21 08:28:01 crc kubenswrapper[4696]: I0321 08:28:01.656746 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1ca3ae85153244eaf6968213d071999f1e13bafc4c31463a015b7912bd5a285d"} Mar 21 08:28:01 crc kubenswrapper[4696]: I0321 08:28:01.656930 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:01 crc kubenswrapper[4696]: I0321 08:28:01.656971 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:01 crc kubenswrapper[4696]: I0321 08:28:01.657961 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:01 crc kubenswrapper[4696]: I0321 08:28:01.657996 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:01 crc kubenswrapper[4696]: I0321 08:28:01.658009 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:01 crc kubenswrapper[4696]: I0321 08:28:01.658384 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:01 crc kubenswrapper[4696]: I0321 08:28:01.658429 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:01 crc kubenswrapper[4696]: I0321 08:28:01.658446 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:01 crc kubenswrapper[4696]: I0321 08:28:01.672028 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 21 08:28:02 crc kubenswrapper[4696]: I0321 08:28:02.481298 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:02Z is after 2026-02-23T05:33:13Z Mar 21 08:28:02 crc kubenswrapper[4696]: I0321 08:28:02.664665 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 21 08:28:02 crc kubenswrapper[4696]: I0321 08:28:02.665543 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 21 08:28:02 crc kubenswrapper[4696]: I0321 08:28:02.668936 4696 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1ca3ae85153244eaf6968213d071999f1e13bafc4c31463a015b7912bd5a285d" exitCode=255 Mar 21 08:28:02 crc kubenswrapper[4696]: I0321 08:28:02.669074 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"1ca3ae85153244eaf6968213d071999f1e13bafc4c31463a015b7912bd5a285d"} Mar 21 08:28:02 crc kubenswrapper[4696]: I0321 08:28:02.669131 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:02 crc kubenswrapper[4696]: I0321 08:28:02.669234 4696 scope.go:117] "RemoveContainer" containerID="573be38366713e9e429e2d24bef5f6c0f923de7715311fc10dfcae6a0e4a4c00" Mar 21 08:28:02 crc kubenswrapper[4696]: I0321 08:28:02.669483 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:02 crc kubenswrapper[4696]: I0321 08:28:02.670323 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:02 crc kubenswrapper[4696]: I0321 08:28:02.670376 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:02 crc kubenswrapper[4696]: I0321 08:28:02.670396 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:02 crc kubenswrapper[4696]: I0321 08:28:02.670990 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:02 crc kubenswrapper[4696]: I0321 08:28:02.671020 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:02 crc kubenswrapper[4696]: I0321 08:28:02.671033 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:02 crc kubenswrapper[4696]: I0321 08:28:02.671474 4696 scope.go:117] "RemoveContainer" containerID="1ca3ae85153244eaf6968213d071999f1e13bafc4c31463a015b7912bd5a285d" Mar 21 08:28:02 crc kubenswrapper[4696]: E0321 08:28:02.671663 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 08:28:03 crc kubenswrapper[4696]: I0321 08:28:03.479288 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:03Z is after 2026-02-23T05:33:13Z Mar 21 08:28:03 crc kubenswrapper[4696]: I0321 08:28:03.674733 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 21 08:28:04 crc kubenswrapper[4696]: I0321 08:28:04.480391 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:04Z is after 2026-02-23T05:33:13Z Mar 21 08:28:04 crc kubenswrapper[4696]: I0321 08:28:04.897872 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:28:04 crc kubenswrapper[4696]: I0321 08:28:04.898194 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:04 crc kubenswrapper[4696]: I0321 08:28:04.899734 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:04 crc kubenswrapper[4696]: I0321 08:28:04.899801 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:04 crc kubenswrapper[4696]: I0321 08:28:04.899874 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:04 crc kubenswrapper[4696]: I0321 08:28:04.900857 4696 scope.go:117] "RemoveContainer" containerID="1ca3ae85153244eaf6968213d071999f1e13bafc4c31463a015b7912bd5a285d" Mar 21 08:28:04 crc kubenswrapper[4696]: E0321 08:28:04.901187 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 08:28:04 crc kubenswrapper[4696]: I0321 08:28:04.903737 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:28:04 crc kubenswrapper[4696]: W0321 08:28:04.980602 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:04Z is after 2026-02-23T05:33:13Z Mar 21 08:28:04 crc kubenswrapper[4696]: E0321 08:28:04.980709 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:04Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 21 08:28:05 crc kubenswrapper[4696]: W0321 08:28:05.045432 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:05Z is after 2026-02-23T05:33:13Z Mar 21 08:28:05 crc kubenswrapper[4696]: E0321 08:28:05.045565 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:05Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 21 08:28:05 crc kubenswrapper[4696]: I0321 08:28:05.140922 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:28:05 crc kubenswrapper[4696]: I0321 08:28:05.141075 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:05 crc kubenswrapper[4696]: I0321 08:28:05.142845 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:05 crc kubenswrapper[4696]: I0321 08:28:05.142912 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:05 crc kubenswrapper[4696]: I0321 08:28:05.142923 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:05 crc kubenswrapper[4696]: I0321 08:28:05.482336 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:05Z is after 2026-02-23T05:33:13Z Mar 21 08:28:05 crc kubenswrapper[4696]: I0321 08:28:05.684213 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:05 crc kubenswrapper[4696]: I0321 08:28:05.685702 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:05 crc kubenswrapper[4696]: I0321 08:28:05.685771 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:05 crc kubenswrapper[4696]: I0321 08:28:05.685791 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:05 crc kubenswrapper[4696]: I0321 08:28:05.686887 4696 scope.go:117] "RemoveContainer" containerID="1ca3ae85153244eaf6968213d071999f1e13bafc4c31463a015b7912bd5a285d" Mar 21 08:28:05 crc kubenswrapper[4696]: E0321 08:28:05.687166 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 08:28:06 crc kubenswrapper[4696]: W0321 08:28:06.104953 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:06Z is after 2026-02-23T05:33:13Z Mar 21 08:28:06 crc kubenswrapper[4696]: E0321 08:28:06.105078 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:28:06Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 21 08:28:06 crc kubenswrapper[4696]: I0321 08:28:06.481387 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:06 crc kubenswrapper[4696]: E0321 08:28:06.609231 4696 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 08:28:06 crc kubenswrapper[4696]: I0321 08:28:06.787516 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:06 crc kubenswrapper[4696]: I0321 08:28:06.789583 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:06 crc kubenswrapper[4696]: I0321 08:28:06.789642 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:06 crc kubenswrapper[4696]: I0321 08:28:06.789656 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:06 crc kubenswrapper[4696]: I0321 08:28:06.789700 4696 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 08:28:06 crc kubenswrapper[4696]: E0321 08:28:06.795723 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 21 08:28:06 crc kubenswrapper[4696]: E0321 08:28:06.797142 4696 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 21 08:28:07 crc kubenswrapper[4696]: I0321 08:28:07.482992 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:07 crc kubenswrapper[4696]: I0321 08:28:07.829481 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:28:07 crc kubenswrapper[4696]: I0321 08:28:07.829691 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:07 crc kubenswrapper[4696]: I0321 08:28:07.831014 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:07 crc kubenswrapper[4696]: I0321 08:28:07.831086 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:07 crc kubenswrapper[4696]: I0321 08:28:07.831111 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:07 crc kubenswrapper[4696]: I0321 08:28:07.832159 4696 scope.go:117] "RemoveContainer" containerID="1ca3ae85153244eaf6968213d071999f1e13bafc4c31463a015b7912bd5a285d" Mar 21 08:28:07 crc kubenswrapper[4696]: E0321 08:28:07.832492 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 08:28:08 crc kubenswrapper[4696]: I0321 08:28:08.485569 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:08 crc kubenswrapper[4696]: I0321 08:28:08.557888 4696 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 21 08:28:08 crc kubenswrapper[4696]: I0321 08:28:08.577886 4696 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 21 08:28:09 crc kubenswrapper[4696]: I0321 08:28:09.037962 4696 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 21 08:28:09 crc kubenswrapper[4696]: I0321 08:28:09.038185 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 21 08:28:09 crc kubenswrapper[4696]: I0321 08:28:09.482660 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.399084 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde04ca5f39c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.472866716 +0000 UTC m=+0.593747449,LastTimestamp:2026-03-21 08:27:46.472866716 +0000 UTC m=+0.593747449,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.404623 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e1deb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539462123 +0000 UTC m=+0.660342836,LastTimestamp:2026-03-21 08:27:46.539462123 +0000 UTC m=+0.660342836,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.408855 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e4d30 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539474224 +0000 UTC m=+0.660354937,LastTimestamp:2026-03-21 08:27:46.539474224 +0000 UTC m=+0.660354937,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.412920 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e6cfd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539482365 +0000 UTC m=+0.660363078,LastTimestamp:2026-03-21 08:27:46.539482365 +0000 UTC m=+0.660363078,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.417048 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0549a23b5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.606310325 +0000 UTC m=+0.727191098,LastTimestamp:2026-03-21 08:27:46.606310325 +0000 UTC m=+0.727191098,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.420237 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ecde0509e1deb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e1deb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539462123 +0000 UTC m=+0.660342836,LastTimestamp:2026-03-21 08:27:46.635170469 +0000 UTC m=+0.756051182,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.422985 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ecde0509e4d30\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e4d30 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539474224 +0000 UTC m=+0.660354937,LastTimestamp:2026-03-21 08:27:46.635188721 +0000 UTC m=+0.756069434,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.426215 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ecde0509e6cfd\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e6cfd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539482365 +0000 UTC m=+0.660363078,LastTimestamp:2026-03-21 08:27:46.635196141 +0000 UTC m=+0.756076854,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.429465 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ecde0509e1deb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e1deb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539462123 +0000 UTC m=+0.660342836,LastTimestamp:2026-03-21 08:27:46.635997468 +0000 UTC m=+0.756878181,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.432908 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ecde0509e4d30\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e4d30 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539474224 +0000 UTC m=+0.660354937,LastTimestamp:2026-03-21 08:27:46.636007648 +0000 UTC m=+0.756888361,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.436571 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ecde0509e6cfd\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e6cfd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539482365 +0000 UTC m=+0.660363078,LastTimestamp:2026-03-21 08:27:46.636015039 +0000 UTC m=+0.756895752,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.439928 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ecde0509e1deb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e1deb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539462123 +0000 UTC m=+0.660342836,LastTimestamp:2026-03-21 08:27:46.637344907 +0000 UTC m=+0.758225620,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.444325 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ecde0509e4d30\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e4d30 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539474224 +0000 UTC m=+0.660354937,LastTimestamp:2026-03-21 08:27:46.637353877 +0000 UTC m=+0.758234590,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.448439 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ecde0509e6cfd\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e6cfd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539482365 +0000 UTC m=+0.660363078,LastTimestamp:2026-03-21 08:27:46.637361978 +0000 UTC m=+0.758242691,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.454276 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ecde0509e1deb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e1deb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539462123 +0000 UTC m=+0.660342836,LastTimestamp:2026-03-21 08:27:46.638135973 +0000 UTC m=+0.759016686,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.458611 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ecde0509e4d30\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e4d30 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539474224 +0000 UTC m=+0.660354937,LastTimestamp:2026-03-21 08:27:46.638148474 +0000 UTC m=+0.759029177,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.464209 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ecde0509e6cfd\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e6cfd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539482365 +0000 UTC m=+0.660363078,LastTimestamp:2026-03-21 08:27:46.638158895 +0000 UTC m=+0.759039608,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.469007 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ecde0509e1deb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e1deb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539462123 +0000 UTC m=+0.660342836,LastTimestamp:2026-03-21 08:27:46.638179587 +0000 UTC m=+0.759060320,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.474045 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ecde0509e4d30\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e4d30 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539474224 +0000 UTC m=+0.660354937,LastTimestamp:2026-03-21 08:27:46.638198499 +0000 UTC m=+0.759079232,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: I0321 08:28:10.478843 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.478772 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ecde0509e6cfd\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e6cfd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539482365 +0000 UTC m=+0.660363078,LastTimestamp:2026-03-21 08:27:46.63821414 +0000 UTC m=+0.759094863,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.480243 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ecde0509e1deb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e1deb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539462123 +0000 UTC m=+0.660342836,LastTimestamp:2026-03-21 08:27:46.638301198 +0000 UTC m=+0.759181911,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.483920 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ecde0509e4d30\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e4d30 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539474224 +0000 UTC m=+0.660354937,LastTimestamp:2026-03-21 08:27:46.638309029 +0000 UTC m=+0.759189742,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.491978 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ecde0509e6cfd\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e6cfd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539482365 +0000 UTC m=+0.660363078,LastTimestamp:2026-03-21 08:27:46.63831585 +0000 UTC m=+0.759196563,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.496668 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ecde0509e1deb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e1deb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539462123 +0000 UTC m=+0.660342836,LastTimestamp:2026-03-21 08:27:46.63841831 +0000 UTC m=+0.759299023,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.502427 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189ecde0509e4d30\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189ecde0509e4d30 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:46.539474224 +0000 UTC m=+0.660354937,LastTimestamp:2026-03-21 08:27:46.63842734 +0000 UTC m=+0.759308053,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.508539 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde06e0c5140 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.033223488 +0000 UTC m=+1.154104211,LastTimestamp:2026-03-21 08:27:47.033223488 +0000 UTC m=+1.154104211,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.512458 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189ecde06e47e9fc openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.037129212 +0000 UTC m=+1.158009925,LastTimestamp:2026-03-21 08:27:47.037129212 +0000 UTC m=+1.158009925,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.516810 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde06e79d7d5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.040401365 +0000 UTC m=+1.161282098,LastTimestamp:2026-03-21 08:27:47.040401365 +0000 UTC m=+1.161282098,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.521486 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ecde0700e6b50 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.066915664 +0000 UTC m=+1.187796427,LastTimestamp:2026-03-21 08:27:47.066915664 +0000 UTC m=+1.187796427,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.522776 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ecde0701f0e0b openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.068005899 +0000 UTC m=+1.188886652,LastTimestamp:2026-03-21 08:27:47.068005899 +0000 UTC m=+1.188886652,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.526608 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ecde091c8f6de openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.632789214 +0000 UTC m=+1.753669927,LastTimestamp:2026-03-21 08:27:47.632789214 +0000 UTC m=+1.753669927,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.531872 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ecde091c9487a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.632810106 +0000 UTC m=+1.753690859,LastTimestamp:2026-03-21 08:27:47.632810106 +0000 UTC m=+1.753690859,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.537300 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde091cbec16 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.632983062 +0000 UTC m=+1.753863785,LastTimestamp:2026-03-21 08:27:47.632983062 +0000 UTC m=+1.753863785,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.543229 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde0922c0078 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.639279736 +0000 UTC m=+1.760160449,LastTimestamp:2026-03-21 08:27:47.639279736 +0000 UTC m=+1.760160449,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.548215 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ecde09265fb2e openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.64307947 +0000 UTC m=+1.763960223,LastTimestamp:2026-03-21 08:27:47.64307947 +0000 UTC m=+1.763960223,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.554595 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189ecde092805ed3 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.644808915 +0000 UTC m=+1.765689638,LastTimestamp:2026-03-21 08:27:47.644808915 +0000 UTC m=+1.765689638,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.558971 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ecde0928f7dc7 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.645799879 +0000 UTC m=+1.766680593,LastTimestamp:2026-03-21 08:27:47.645799879 +0000 UTC m=+1.766680593,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.562933 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ecde092a4cf19 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.647196953 +0000 UTC m=+1.768077686,LastTimestamp:2026-03-21 08:27:47.647196953 +0000 UTC m=+1.768077686,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.567737 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde092b5b74a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.64830497 +0000 UTC m=+1.769185683,LastTimestamp:2026-03-21 08:27:47.64830497 +0000 UTC m=+1.769185683,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.572528 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde09332f314 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.656512276 +0000 UTC m=+1.777393009,LastTimestamp:2026-03-21 08:27:47.656512276 +0000 UTC m=+1.777393009,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.577773 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189ecde093ee542a openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.668792362 +0000 UTC m=+1.789673075,LastTimestamp:2026-03-21 08:27:47.668792362 +0000 UTC m=+1.789673075,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.584542 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ecde0a50e0f94 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.956084628 +0000 UTC m=+2.076965381,LastTimestamp:2026-03-21 08:27:47.956084628 +0000 UTC m=+2.076965381,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.590007 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ecde0a5f0f981 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.970955649 +0000 UTC m=+2.091836412,LastTimestamp:2026-03-21 08:27:47.970955649 +0000 UTC m=+2.091836412,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.594205 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ecde0a60f268e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.972933262 +0000 UTC m=+2.093814015,LastTimestamp:2026-03-21 08:27:47.972933262 +0000 UTC m=+2.093814015,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.600313 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ecde0b2369022 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:48.176842786 +0000 UTC m=+2.297723509,LastTimestamp:2026-03-21 08:27:48.176842786 +0000 UTC m=+2.297723509,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.604785 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ecde0b33e1c8a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:48.194114698 +0000 UTC m=+2.314995411,LastTimestamp:2026-03-21 08:27:48.194114698 +0000 UTC m=+2.314995411,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.609737 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ecde0b3548065 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:48.195582053 +0000 UTC m=+2.316462776,LastTimestamp:2026-03-21 08:27:48.195582053 +0000 UTC m=+2.316462776,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.614833 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ecde0bf6aec77 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:48.398378103 +0000 UTC m=+2.519258816,LastTimestamp:2026-03-21 08:27:48.398378103 +0000 UTC m=+2.519258816,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: W0321 08:28:10.615426 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.615472 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.618699 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ecde0c0179d5e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:48.409695582 +0000 UTC m=+2.530576295,LastTimestamp:2026-03-21 08:27:48.409695582 +0000 UTC m=+2.530576295,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.623700 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ecde0c8858347 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:48.551115591 +0000 UTC m=+2.671996344,LastTimestamp:2026-03-21 08:27:48.551115591 +0000 UTC m=+2.671996344,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.629926 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde0c88d104d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:48.551610445 +0000 UTC m=+2.672491168,LastTimestamp:2026-03-21 08:27:48.551610445 +0000 UTC m=+2.672491168,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.633536 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189ecde0c8d7068f openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:48.556457615 +0000 UTC m=+2.677338328,LastTimestamp:2026-03-21 08:27:48.556457615 +0000 UTC m=+2.677338328,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.639169 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde0c98615ed openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:48.567930349 +0000 UTC m=+2.688811062,LastTimestamp:2026-03-21 08:27:48.567930349 +0000 UTC m=+2.688811062,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.646293 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde0d5b9409f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:48.772610207 +0000 UTC m=+2.893490920,LastTimestamp:2026-03-21 08:27:48.772610207 +0000 UTC m=+2.893490920,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.650953 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189ecde0d5bd6f83 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:48.772884355 +0000 UTC m=+2.893765068,LastTimestamp:2026-03-21 08:27:48.772884355 +0000 UTC m=+2.893765068,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.657220 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ecde0d5befe83 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:48.772986499 +0000 UTC m=+2.893867212,LastTimestamp:2026-03-21 08:27:48.772986499 +0000 UTC m=+2.893867212,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.662362 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde0d66b3cff openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:48.784274687 +0000 UTC m=+2.905155400,LastTimestamp:2026-03-21 08:27:48.784274687 +0000 UTC m=+2.905155400,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.669072 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde0d681736c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:48.785730412 +0000 UTC m=+2.906611125,LastTimestamp:2026-03-21 08:27:48.785730412 +0000 UTC m=+2.906611125,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.672943 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ecde0d681dc81 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:48.785757313 +0000 UTC m=+2.906638026,LastTimestamp:2026-03-21 08:27:48.785757313 +0000 UTC m=+2.906638026,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.677560 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde0d69a2bdc openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:48.787350492 +0000 UTC m=+2.908231195,LastTimestamp:2026-03-21 08:27:48.787350492 +0000 UTC m=+2.908231195,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.681726 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ecde0d6a1e4b7 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:48.787856567 +0000 UTC m=+2.908737280,LastTimestamp:2026-03-21 08:27:48.787856567 +0000 UTC m=+2.908737280,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.688402 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189ecde0d6f6a2a0 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:48.793410208 +0000 UTC m=+2.914290921,LastTimestamp:2026-03-21 08:27:48.793410208 +0000 UTC m=+2.914290921,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.694219 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde0d8190a53 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:48.812442195 +0000 UTC m=+2.933322908,LastTimestamp:2026-03-21 08:27:48.812442195 +0000 UTC m=+2.933322908,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.700201 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ecde0e3c5a430 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.008303152 +0000 UTC m=+3.129183905,LastTimestamp:2026-03-21 08:27:49.008303152 +0000 UTC m=+3.129183905,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.705089 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde0e3e2cbee openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.01021387 +0000 UTC m=+3.131094593,LastTimestamp:2026-03-21 08:27:49.01021387 +0000 UTC m=+3.131094593,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.709268 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ecde0e4792ebc openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.020069564 +0000 UTC m=+3.140950307,LastTimestamp:2026-03-21 08:27:49.020069564 +0000 UTC m=+3.140950307,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.712854 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde0e485b411 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.020890129 +0000 UTC m=+3.141770883,LastTimestamp:2026-03-21 08:27:49.020890129 +0000 UTC m=+3.141770883,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.716593 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ecde0e49165f3 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.021656563 +0000 UTC m=+3.142537266,LastTimestamp:2026-03-21 08:27:49.021656563 +0000 UTC m=+3.142537266,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.721255 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde0e49dea8a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.022476938 +0000 UTC m=+3.143357681,LastTimestamp:2026-03-21 08:27:49.022476938 +0000 UTC m=+3.143357681,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.725433 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ecde0f0bfdc13 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.226028051 +0000 UTC m=+3.346908764,LastTimestamp:2026-03-21 08:27:49.226028051 +0000 UTC m=+3.346908764,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.729948 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde0f0d6948b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.227517067 +0000 UTC m=+3.348397780,LastTimestamp:2026-03-21 08:27:49.227517067 +0000 UTC m=+3.348397780,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.735708 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189ecde0f15f9d50 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.236497744 +0000 UTC m=+3.357378457,LastTimestamp:2026-03-21 08:27:49.236497744 +0000 UTC m=+3.357378457,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.739734 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde0f1f36ed4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.246185172 +0000 UTC m=+3.367065885,LastTimestamp:2026-03-21 08:27:49.246185172 +0000 UTC m=+3.367065885,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.745840 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde0f205d7a2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.24739165 +0000 UTC m=+3.368272373,LastTimestamp:2026-03-21 08:27:49.24739165 +0000 UTC m=+3.368272373,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.752409 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde0fb946790 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.407729552 +0000 UTC m=+3.528610265,LastTimestamp:2026-03-21 08:27:49.407729552 +0000 UTC m=+3.528610265,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.756282 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde0fc46db46 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.419424582 +0000 UTC m=+3.540305295,LastTimestamp:2026-03-21 08:27:49.419424582 +0000 UTC m=+3.540305295,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.762146 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde0fc6081b9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.421105593 +0000 UTC m=+3.541986306,LastTimestamp:2026-03-21 08:27:49.421105593 +0000 UTC m=+3.541986306,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.772635 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde1059a7ea3 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.575900835 +0000 UTC m=+3.696781548,LastTimestamp:2026-03-21 08:27:49.575900835 +0000 UTC m=+3.696781548,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.778107 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde10804e088 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.616427144 +0000 UTC m=+3.737307857,LastTimestamp:2026-03-21 08:27:49.616427144 +0000 UTC m=+3.737307857,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.784514 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde108aea2c9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.627552457 +0000 UTC m=+3.748433190,LastTimestamp:2026-03-21 08:27:49.627552457 +0000 UTC m=+3.748433190,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.790723 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde1126115a2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.79024221 +0000 UTC m=+3.911122923,LastTimestamp:2026-03-21 08:27:49.79024221 +0000 UTC m=+3.911122923,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.798005 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde112e89cc4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.799124164 +0000 UTC m=+3.920004877,LastTimestamp:2026-03-21 08:27:49.799124164 +0000 UTC m=+3.920004877,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.804673 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde14332adc0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:50.609284544 +0000 UTC m=+4.730165257,LastTimestamp:2026-03-21 08:27:50.609284544 +0000 UTC m=+4.730165257,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.809929 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde150892ab9 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:50.833056441 +0000 UTC m=+4.953937194,LastTimestamp:2026-03-21 08:27:50.833056441 +0000 UTC m=+4.953937194,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.815862 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde15116b10e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:50.842331406 +0000 UTC m=+4.963212119,LastTimestamp:2026-03-21 08:27:50.842331406 +0000 UTC m=+4.963212119,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.820635 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde1512759f4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:50.84342322 +0000 UTC m=+4.964303923,LastTimestamp:2026-03-21 08:27:50.84342322 +0000 UTC m=+4.964303923,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.825433 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde15bf03852 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:51.024359506 +0000 UTC m=+5.145240219,LastTimestamp:2026-03-21 08:27:51.024359506 +0000 UTC m=+5.145240219,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.830390 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde15ca5beb3 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:51.036255923 +0000 UTC m=+5.157136636,LastTimestamp:2026-03-21 08:27:51.036255923 +0000 UTC m=+5.157136636,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.835744 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde15cb3dca4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:51.037181092 +0000 UTC m=+5.158061805,LastTimestamp:2026-03-21 08:27:51.037181092 +0000 UTC m=+5.158061805,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.842146 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde1690e42b1 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:51.244432049 +0000 UTC m=+5.365312762,LastTimestamp:2026-03-21 08:27:51.244432049 +0000 UTC m=+5.365312762,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.845077 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde169a426a8 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:51.254255272 +0000 UTC m=+5.375135985,LastTimestamp:2026-03-21 08:27:51.254255272 +0000 UTC m=+5.375135985,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.850707 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde169b45e30 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:51.255318064 +0000 UTC m=+5.376198777,LastTimestamp:2026-03-21 08:27:51.255318064 +0000 UTC m=+5.376198777,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.857548 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde17588e076 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:51.453794422 +0000 UTC m=+5.574675145,LastTimestamp:2026-03-21 08:27:51.453794422 +0000 UTC m=+5.574675145,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.861885 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde17649bb29 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:51.466433321 +0000 UTC m=+5.587314034,LastTimestamp:2026-03-21 08:27:51.466433321 +0000 UTC m=+5.587314034,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.866188 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde1765bd77a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:51.467620218 +0000 UTC m=+5.588500941,LastTimestamp:2026-03-21 08:27:51.467620218 +0000 UTC m=+5.588500941,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.871230 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde188cf53be openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:51.777178558 +0000 UTC m=+5.898059261,LastTimestamp:2026-03-21 08:27:51.777178558 +0000 UTC m=+5.898059261,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.875265 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189ecde18a09dd5e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:51.797792094 +0000 UTC m=+5.918672807,LastTimestamp:2026-03-21 08:27:51.797792094 +0000 UTC m=+5.918672807,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.884856 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 21 08:28:10 crc kubenswrapper[4696]: &Event{ObjectMeta:{kube-controller-manager-crc.189ecde339962e45 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 21 08:28:10 crc kubenswrapper[4696]: body: Mar 21 08:28:10 crc kubenswrapper[4696]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:59.037967941 +0000 UTC m=+13.158848644,LastTimestamp:2026-03-21 08:27:59.037967941 +0000 UTC m=+13.158848644,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 21 08:28:10 crc kubenswrapper[4696]: > Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.890971 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ecde33996ee76 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:59.038017142 +0000 UTC m=+13.158897855,LastTimestamp:2026-03-21 08:27:59.038017142 +0000 UTC m=+13.158897855,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.895258 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 21 08:28:10 crc kubenswrapper[4696]: &Event{ObjectMeta:{kube-apiserver-crc.189ecde38aa328ff openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 21 08:28:10 crc kubenswrapper[4696]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 21 08:28:10 crc kubenswrapper[4696]: Mar 21 08:28:10 crc kubenswrapper[4696]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:28:00.397773055 +0000 UTC m=+14.518653788,LastTimestamp:2026-03-21 08:28:00.397773055 +0000 UTC m=+14.518653788,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 21 08:28:10 crc kubenswrapper[4696]: > Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.900622 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde38aa47a31 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:28:00.397859377 +0000 UTC m=+14.518740100,LastTimestamp:2026-03-21 08:28:00.397859377 +0000 UTC m=+14.518740100,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.908054 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189ecde38aa328ff\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 21 08:28:10 crc kubenswrapper[4696]: &Event{ObjectMeta:{kube-apiserver-crc.189ecde38aa328ff openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 21 08:28:10 crc kubenswrapper[4696]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 21 08:28:10 crc kubenswrapper[4696]: Mar 21 08:28:10 crc kubenswrapper[4696]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:28:00.397773055 +0000 UTC m=+14.518653788,LastTimestamp:2026-03-21 08:28:00.404304543 +0000 UTC m=+14.525185266,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 21 08:28:10 crc kubenswrapper[4696]: > Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.915485 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189ecde38aa47a31\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde38aa47a31 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:28:00.397859377 +0000 UTC m=+14.518740100,LastTimestamp:2026-03-21 08:28:00.404362705 +0000 UTC m=+14.525243428,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.920380 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189ecde0fc6081b9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde0fc6081b9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.421105593 +0000 UTC m=+3.541986306,LastTimestamp:2026-03-21 08:28:00.651573945 +0000 UTC m=+14.772454658,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.925048 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189ecde10804e088\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde10804e088 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.616427144 +0000 UTC m=+3.737307857,LastTimestamp:2026-03-21 08:28:00.815867235 +0000 UTC m=+14.936747948,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.931338 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189ecde108aea2c9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189ecde108aea2c9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:49.627552457 +0000 UTC m=+3.748433190,LastTimestamp:2026-03-21 08:28:00.829357232 +0000 UTC m=+14.950237965,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.938303 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 21 08:28:10 crc kubenswrapper[4696]: &Event{ObjectMeta:{kube-controller-manager-crc.189ecde58da45336 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 21 08:28:10 crc kubenswrapper[4696]: body: Mar 21 08:28:10 crc kubenswrapper[4696]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:28:09.038115638 +0000 UTC m=+23.158996381,LastTimestamp:2026-03-21 08:28:09.038115638 +0000 UTC m=+23.158996381,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 21 08:28:10 crc kubenswrapper[4696]: > Mar 21 08:28:10 crc kubenswrapper[4696]: E0321 08:28:10.943709 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ecde58da64063 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:28:09.038241891 +0000 UTC m=+23.159122634,LastTimestamp:2026-03-21 08:28:09.038241891 +0000 UTC m=+23.159122634,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:11 crc kubenswrapper[4696]: I0321 08:28:11.215369 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:28:11 crc kubenswrapper[4696]: I0321 08:28:11.216414 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:11 crc kubenswrapper[4696]: I0321 08:28:11.219387 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:11 crc kubenswrapper[4696]: I0321 08:28:11.219598 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:11 crc kubenswrapper[4696]: I0321 08:28:11.219755 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:11 crc kubenswrapper[4696]: I0321 08:28:11.220973 4696 scope.go:117] "RemoveContainer" containerID="1ca3ae85153244eaf6968213d071999f1e13bafc4c31463a015b7912bd5a285d" Mar 21 08:28:11 crc kubenswrapper[4696]: E0321 08:28:11.221552 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 08:28:11 crc kubenswrapper[4696]: I0321 08:28:11.483159 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:12 crc kubenswrapper[4696]: W0321 08:28:12.277975 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:12 crc kubenswrapper[4696]: E0321 08:28:12.278080 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 21 08:28:12 crc kubenswrapper[4696]: I0321 08:28:12.483612 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:13 crc kubenswrapper[4696]: W0321 08:28:13.308547 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 21 08:28:13 crc kubenswrapper[4696]: E0321 08:28:13.308658 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 21 08:28:13 crc kubenswrapper[4696]: I0321 08:28:13.480258 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:13 crc kubenswrapper[4696]: I0321 08:28:13.797236 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:13 crc kubenswrapper[4696]: I0321 08:28:13.799345 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:13 crc kubenswrapper[4696]: I0321 08:28:13.799403 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:13 crc kubenswrapper[4696]: I0321 08:28:13.799418 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:13 crc kubenswrapper[4696]: I0321 08:28:13.799448 4696 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 08:28:13 crc kubenswrapper[4696]: E0321 08:28:13.804264 4696 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 21 08:28:13 crc kubenswrapper[4696]: E0321 08:28:13.804904 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 21 08:28:14 crc kubenswrapper[4696]: W0321 08:28:14.369167 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 21 08:28:14 crc kubenswrapper[4696]: E0321 08:28:14.369290 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 21 08:28:14 crc kubenswrapper[4696]: I0321 08:28:14.483950 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:15 crc kubenswrapper[4696]: I0321 08:28:15.482369 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:16 crc kubenswrapper[4696]: I0321 08:28:16.480131 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:16 crc kubenswrapper[4696]: E0321 08:28:16.609498 4696 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 08:28:17 crc kubenswrapper[4696]: I0321 08:28:17.483401 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:18 crc kubenswrapper[4696]: I0321 08:28:18.484609 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:19 crc kubenswrapper[4696]: I0321 08:28:19.016019 4696 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:59280->192.168.126.11:10357: read: connection reset by peer" start-of-body= Mar 21 08:28:19 crc kubenswrapper[4696]: I0321 08:28:19.016147 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:59280->192.168.126.11:10357: read: connection reset by peer" Mar 21 08:28:19 crc kubenswrapper[4696]: I0321 08:28:19.016245 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:28:19 crc kubenswrapper[4696]: I0321 08:28:19.016477 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:19 crc kubenswrapper[4696]: I0321 08:28:19.018410 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:19 crc kubenswrapper[4696]: I0321 08:28:19.018469 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:19 crc kubenswrapper[4696]: I0321 08:28:19.018488 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:19 crc kubenswrapper[4696]: I0321 08:28:19.019218 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"6f9f2c48d413fd10728233381f94e3fb94d5b73000d1337244a76a895b2bcb31"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 21 08:28:19 crc kubenswrapper[4696]: I0321 08:28:19.019481 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://6f9f2c48d413fd10728233381f94e3fb94d5b73000d1337244a76a895b2bcb31" gracePeriod=30 Mar 21 08:28:19 crc kubenswrapper[4696]: E0321 08:28:19.026252 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 21 08:28:19 crc kubenswrapper[4696]: &Event{ObjectMeta:{kube-controller-manager-crc.189ecde7e0606f12 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": read tcp 192.168.126.11:59280->192.168.126.11:10357: read: connection reset by peer Mar 21 08:28:19 crc kubenswrapper[4696]: body: Mar 21 08:28:19 crc kubenswrapper[4696]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:28:19.016109842 +0000 UTC m=+33.136990595,LastTimestamp:2026-03-21 08:28:19.016109842 +0000 UTC m=+33.136990595,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 21 08:28:19 crc kubenswrapper[4696]: > Mar 21 08:28:19 crc kubenswrapper[4696]: E0321 08:28:19.035363 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ecde7e061ce0f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:59280->192.168.126.11:10357: read: connection reset by peer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:28:19.016199695 +0000 UTC m=+33.137080448,LastTimestamp:2026-03-21 08:28:19.016199695 +0000 UTC m=+33.137080448,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:19 crc kubenswrapper[4696]: E0321 08:28:19.044106 4696 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ecde7e0935c93 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:28:19.019447443 +0000 UTC m=+33.140328196,LastTimestamp:2026-03-21 08:28:19.019447443 +0000 UTC m=+33.140328196,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:19 crc kubenswrapper[4696]: E0321 08:28:19.052254 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189ecde092a4cf19\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ecde092a4cf19 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.647196953 +0000 UTC m=+1.768077686,LastTimestamp:2026-03-21 08:28:19.047922659 +0000 UTC m=+33.168803392,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:19 crc kubenswrapper[4696]: E0321 08:28:19.272231 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189ecde0a50e0f94\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ecde0a50e0f94 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.956084628 +0000 UTC m=+2.076965381,LastTimestamp:2026-03-21 08:28:19.266231311 +0000 UTC m=+33.387112054,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:19 crc kubenswrapper[4696]: E0321 08:28:19.281566 4696 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189ecde0a5f0f981\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189ecde0a5f0f981 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:27:47.970955649 +0000 UTC m=+2.091836412,LastTimestamp:2026-03-21 08:28:19.279277607 +0000 UTC m=+33.400158350,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:28:19 crc kubenswrapper[4696]: I0321 08:28:19.485162 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:19 crc kubenswrapper[4696]: I0321 08:28:19.733046 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 21 08:28:19 crc kubenswrapper[4696]: I0321 08:28:19.733726 4696 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="6f9f2c48d413fd10728233381f94e3fb94d5b73000d1337244a76a895b2bcb31" exitCode=255 Mar 21 08:28:19 crc kubenswrapper[4696]: I0321 08:28:19.733782 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"6f9f2c48d413fd10728233381f94e3fb94d5b73000d1337244a76a895b2bcb31"} Mar 21 08:28:19 crc kubenswrapper[4696]: I0321 08:28:19.733810 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"db7407d284c4174177aebe356467c28d58b73b727257f252a30640c3b6b3af09"} Mar 21 08:28:19 crc kubenswrapper[4696]: I0321 08:28:19.733963 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:19 crc kubenswrapper[4696]: I0321 08:28:19.735202 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:19 crc kubenswrapper[4696]: I0321 08:28:19.735257 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:19 crc kubenswrapper[4696]: I0321 08:28:19.735273 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:20 crc kubenswrapper[4696]: I0321 08:28:20.481637 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:20 crc kubenswrapper[4696]: I0321 08:28:20.804647 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:20 crc kubenswrapper[4696]: I0321 08:28:20.806163 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:20 crc kubenswrapper[4696]: I0321 08:28:20.806211 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:20 crc kubenswrapper[4696]: I0321 08:28:20.806229 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:20 crc kubenswrapper[4696]: I0321 08:28:20.806262 4696 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 08:28:20 crc kubenswrapper[4696]: E0321 08:28:20.810509 4696 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 21 08:28:20 crc kubenswrapper[4696]: E0321 08:28:20.811066 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 21 08:28:21 crc kubenswrapper[4696]: I0321 08:28:21.482163 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:22 crc kubenswrapper[4696]: I0321 08:28:22.478204 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:23 crc kubenswrapper[4696]: I0321 08:28:23.481027 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:23 crc kubenswrapper[4696]: I0321 08:28:23.534010 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:23 crc kubenswrapper[4696]: I0321 08:28:23.535956 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:23 crc kubenswrapper[4696]: I0321 08:28:23.536032 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:23 crc kubenswrapper[4696]: I0321 08:28:23.536060 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:23 crc kubenswrapper[4696]: I0321 08:28:23.537183 4696 scope.go:117] "RemoveContainer" containerID="1ca3ae85153244eaf6968213d071999f1e13bafc4c31463a015b7912bd5a285d" Mar 21 08:28:23 crc kubenswrapper[4696]: I0321 08:28:23.769952 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:28:23 crc kubenswrapper[4696]: I0321 08:28:23.770261 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:23 crc kubenswrapper[4696]: I0321 08:28:23.771691 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:23 crc kubenswrapper[4696]: I0321 08:28:23.771735 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:23 crc kubenswrapper[4696]: I0321 08:28:23.771748 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:24 crc kubenswrapper[4696]: I0321 08:28:24.483164 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:24 crc kubenswrapper[4696]: I0321 08:28:24.749311 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 21 08:28:24 crc kubenswrapper[4696]: I0321 08:28:24.749863 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 21 08:28:24 crc kubenswrapper[4696]: I0321 08:28:24.753274 4696 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fb88caa7f0b405bdd9a4a389a796d8fd9b9d18a61262d50671fd3ce11863d98f" exitCode=255 Mar 21 08:28:24 crc kubenswrapper[4696]: I0321 08:28:24.753348 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"fb88caa7f0b405bdd9a4a389a796d8fd9b9d18a61262d50671fd3ce11863d98f"} Mar 21 08:28:24 crc kubenswrapper[4696]: I0321 08:28:24.753417 4696 scope.go:117] "RemoveContainer" containerID="1ca3ae85153244eaf6968213d071999f1e13bafc4c31463a015b7912bd5a285d" Mar 21 08:28:24 crc kubenswrapper[4696]: I0321 08:28:24.753623 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:24 crc kubenswrapper[4696]: I0321 08:28:24.754528 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:24 crc kubenswrapper[4696]: I0321 08:28:24.754574 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:24 crc kubenswrapper[4696]: I0321 08:28:24.754587 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:24 crc kubenswrapper[4696]: I0321 08:28:24.755192 4696 scope.go:117] "RemoveContainer" containerID="fb88caa7f0b405bdd9a4a389a796d8fd9b9d18a61262d50671fd3ce11863d98f" Mar 21 08:28:24 crc kubenswrapper[4696]: E0321 08:28:24.755375 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 08:28:25 crc kubenswrapper[4696]: I0321 08:28:25.480024 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:25 crc kubenswrapper[4696]: I0321 08:28:25.757862 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 21 08:28:26 crc kubenswrapper[4696]: I0321 08:28:26.037402 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:28:26 crc kubenswrapper[4696]: I0321 08:28:26.037590 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:26 crc kubenswrapper[4696]: I0321 08:28:26.038956 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:26 crc kubenswrapper[4696]: I0321 08:28:26.039022 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:26 crc kubenswrapper[4696]: I0321 08:28:26.039035 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:26 crc kubenswrapper[4696]: I0321 08:28:26.044880 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:28:26 crc kubenswrapper[4696]: I0321 08:28:26.479519 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:26 crc kubenswrapper[4696]: E0321 08:28:26.610765 4696 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 08:28:26 crc kubenswrapper[4696]: I0321 08:28:26.762650 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:26 crc kubenswrapper[4696]: I0321 08:28:26.763382 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:26 crc kubenswrapper[4696]: I0321 08:28:26.763409 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:26 crc kubenswrapper[4696]: I0321 08:28:26.763417 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:27 crc kubenswrapper[4696]: I0321 08:28:27.482137 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:27 crc kubenswrapper[4696]: I0321 08:28:27.811441 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:27 crc kubenswrapper[4696]: I0321 08:28:27.812891 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:27 crc kubenswrapper[4696]: I0321 08:28:27.812937 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:27 crc kubenswrapper[4696]: I0321 08:28:27.812949 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:27 crc kubenswrapper[4696]: I0321 08:28:27.812979 4696 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 08:28:27 crc kubenswrapper[4696]: E0321 08:28:27.815454 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 21 08:28:27 crc kubenswrapper[4696]: E0321 08:28:27.815707 4696 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 21 08:28:27 crc kubenswrapper[4696]: I0321 08:28:27.830104 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:28:27 crc kubenswrapper[4696]: I0321 08:28:27.830273 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:27 crc kubenswrapper[4696]: I0321 08:28:27.831384 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:27 crc kubenswrapper[4696]: I0321 08:28:27.831417 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:27 crc kubenswrapper[4696]: I0321 08:28:27.831427 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:27 crc kubenswrapper[4696]: I0321 08:28:27.832014 4696 scope.go:117] "RemoveContainer" containerID="fb88caa7f0b405bdd9a4a389a796d8fd9b9d18a61262d50671fd3ce11863d98f" Mar 21 08:28:27 crc kubenswrapper[4696]: E0321 08:28:27.832182 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 08:28:28 crc kubenswrapper[4696]: I0321 08:28:28.478853 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:29 crc kubenswrapper[4696]: I0321 08:28:29.479072 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:30 crc kubenswrapper[4696]: I0321 08:28:30.481686 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:30 crc kubenswrapper[4696]: W0321 08:28:30.828520 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 21 08:28:30 crc kubenswrapper[4696]: E0321 08:28:30.828565 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 21 08:28:31 crc kubenswrapper[4696]: W0321 08:28:31.173608 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 21 08:28:31 crc kubenswrapper[4696]: E0321 08:28:31.174068 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 21 08:28:31 crc kubenswrapper[4696]: I0321 08:28:31.215513 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:28:31 crc kubenswrapper[4696]: I0321 08:28:31.216125 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:31 crc kubenswrapper[4696]: I0321 08:28:31.217552 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:31 crc kubenswrapper[4696]: I0321 08:28:31.217599 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:31 crc kubenswrapper[4696]: I0321 08:28:31.217611 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:31 crc kubenswrapper[4696]: I0321 08:28:31.221477 4696 scope.go:117] "RemoveContainer" containerID="fb88caa7f0b405bdd9a4a389a796d8fd9b9d18a61262d50671fd3ce11863d98f" Mar 21 08:28:31 crc kubenswrapper[4696]: E0321 08:28:31.221952 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 08:28:31 crc kubenswrapper[4696]: I0321 08:28:31.478890 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:32 crc kubenswrapper[4696]: W0321 08:28:32.207452 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:32 crc kubenswrapper[4696]: E0321 08:28:32.208227 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 21 08:28:32 crc kubenswrapper[4696]: I0321 08:28:32.482214 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:33 crc kubenswrapper[4696]: I0321 08:28:33.479752 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:33 crc kubenswrapper[4696]: I0321 08:28:33.773257 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:28:33 crc kubenswrapper[4696]: I0321 08:28:33.773436 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:33 crc kubenswrapper[4696]: I0321 08:28:33.774451 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:33 crc kubenswrapper[4696]: I0321 08:28:33.774504 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:33 crc kubenswrapper[4696]: I0321 08:28:33.774517 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:34 crc kubenswrapper[4696]: I0321 08:28:34.479807 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:34 crc kubenswrapper[4696]: W0321 08:28:34.729989 4696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 21 08:28:34 crc kubenswrapper[4696]: E0321 08:28:34.730062 4696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 21 08:28:34 crc kubenswrapper[4696]: I0321 08:28:34.816441 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:34 crc kubenswrapper[4696]: I0321 08:28:34.818097 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:34 crc kubenswrapper[4696]: I0321 08:28:34.818132 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:34 crc kubenswrapper[4696]: I0321 08:28:34.818141 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:34 crc kubenswrapper[4696]: I0321 08:28:34.818163 4696 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 08:28:34 crc kubenswrapper[4696]: E0321 08:28:34.823849 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 21 08:28:34 crc kubenswrapper[4696]: E0321 08:28:34.823900 4696 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 21 08:28:35 crc kubenswrapper[4696]: I0321 08:28:35.481019 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:36 crc kubenswrapper[4696]: I0321 08:28:36.481150 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:36 crc kubenswrapper[4696]: E0321 08:28:36.611662 4696 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 08:28:37 crc kubenswrapper[4696]: I0321 08:28:37.482493 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:38 crc kubenswrapper[4696]: I0321 08:28:38.480735 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:39 crc kubenswrapper[4696]: I0321 08:28:39.482358 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:39 crc kubenswrapper[4696]: I0321 08:28:39.988865 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 08:28:39 crc kubenswrapper[4696]: I0321 08:28:39.989118 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:39 crc kubenswrapper[4696]: I0321 08:28:39.990903 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:39 crc kubenswrapper[4696]: I0321 08:28:39.990967 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:39 crc kubenswrapper[4696]: I0321 08:28:39.990988 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:40 crc kubenswrapper[4696]: I0321 08:28:40.482675 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:41 crc kubenswrapper[4696]: I0321 08:28:41.480388 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:41 crc kubenswrapper[4696]: I0321 08:28:41.824996 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:41 crc kubenswrapper[4696]: I0321 08:28:41.827604 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:41 crc kubenswrapper[4696]: I0321 08:28:41.827675 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:41 crc kubenswrapper[4696]: I0321 08:28:41.827689 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:41 crc kubenswrapper[4696]: I0321 08:28:41.827735 4696 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 08:28:41 crc kubenswrapper[4696]: E0321 08:28:41.831946 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 21 08:28:41 crc kubenswrapper[4696]: E0321 08:28:41.832054 4696 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 21 08:28:42 crc kubenswrapper[4696]: I0321 08:28:42.482284 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:43 crc kubenswrapper[4696]: I0321 08:28:43.481134 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:44 crc kubenswrapper[4696]: I0321 08:28:44.481308 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:45 crc kubenswrapper[4696]: I0321 08:28:45.480095 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:45 crc kubenswrapper[4696]: I0321 08:28:45.534322 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:45 crc kubenswrapper[4696]: I0321 08:28:45.535499 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:45 crc kubenswrapper[4696]: I0321 08:28:45.535560 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:45 crc kubenswrapper[4696]: I0321 08:28:45.535572 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:45 crc kubenswrapper[4696]: I0321 08:28:45.536236 4696 scope.go:117] "RemoveContainer" containerID="fb88caa7f0b405bdd9a4a389a796d8fd9b9d18a61262d50671fd3ce11863d98f" Mar 21 08:28:45 crc kubenswrapper[4696]: I0321 08:28:45.811735 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 21 08:28:45 crc kubenswrapper[4696]: I0321 08:28:45.813704 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a78fedb8a70e8b5cec10a972e63964e900746253f6add19cbedf2e47af762244"} Mar 21 08:28:45 crc kubenswrapper[4696]: I0321 08:28:45.813985 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:45 crc kubenswrapper[4696]: I0321 08:28:45.814940 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:45 crc kubenswrapper[4696]: I0321 08:28:45.814983 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:45 crc kubenswrapper[4696]: I0321 08:28:45.815011 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:46 crc kubenswrapper[4696]: I0321 08:28:46.481315 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:46 crc kubenswrapper[4696]: E0321 08:28:46.611978 4696 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 08:28:46 crc kubenswrapper[4696]: I0321 08:28:46.819298 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 21 08:28:46 crc kubenswrapper[4696]: I0321 08:28:46.820466 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 21 08:28:46 crc kubenswrapper[4696]: I0321 08:28:46.823071 4696 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a78fedb8a70e8b5cec10a972e63964e900746253f6add19cbedf2e47af762244" exitCode=255 Mar 21 08:28:46 crc kubenswrapper[4696]: I0321 08:28:46.823150 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a78fedb8a70e8b5cec10a972e63964e900746253f6add19cbedf2e47af762244"} Mar 21 08:28:46 crc kubenswrapper[4696]: I0321 08:28:46.823215 4696 scope.go:117] "RemoveContainer" containerID="fb88caa7f0b405bdd9a4a389a796d8fd9b9d18a61262d50671fd3ce11863d98f" Mar 21 08:28:46 crc kubenswrapper[4696]: I0321 08:28:46.823519 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:46 crc kubenswrapper[4696]: I0321 08:28:46.825096 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:46 crc kubenswrapper[4696]: I0321 08:28:46.825164 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:46 crc kubenswrapper[4696]: I0321 08:28:46.825190 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:46 crc kubenswrapper[4696]: I0321 08:28:46.826589 4696 scope.go:117] "RemoveContainer" containerID="a78fedb8a70e8b5cec10a972e63964e900746253f6add19cbedf2e47af762244" Mar 21 08:28:46 crc kubenswrapper[4696]: E0321 08:28:46.826998 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 08:28:47 crc kubenswrapper[4696]: I0321 08:28:47.479862 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:47 crc kubenswrapper[4696]: I0321 08:28:47.828913 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 21 08:28:47 crc kubenswrapper[4696]: I0321 08:28:47.829275 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:28:47 crc kubenswrapper[4696]: I0321 08:28:47.831490 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:47 crc kubenswrapper[4696]: I0321 08:28:47.832519 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:47 crc kubenswrapper[4696]: I0321 08:28:47.832553 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:47 crc kubenswrapper[4696]: I0321 08:28:47.832595 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:47 crc kubenswrapper[4696]: I0321 08:28:47.833154 4696 scope.go:117] "RemoveContainer" containerID="a78fedb8a70e8b5cec10a972e63964e900746253f6add19cbedf2e47af762244" Mar 21 08:28:47 crc kubenswrapper[4696]: E0321 08:28:47.833348 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 08:28:48 crc kubenswrapper[4696]: I0321 08:28:48.482009 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:48 crc kubenswrapper[4696]: I0321 08:28:48.832164 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:48 crc kubenswrapper[4696]: I0321 08:28:48.835330 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:48 crc kubenswrapper[4696]: I0321 08:28:48.835466 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:48 crc kubenswrapper[4696]: I0321 08:28:48.835556 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:48 crc kubenswrapper[4696]: I0321 08:28:48.835671 4696 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 08:28:48 crc kubenswrapper[4696]: E0321 08:28:48.837538 4696 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 21 08:28:48 crc kubenswrapper[4696]: E0321 08:28:48.838292 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 21 08:28:49 crc kubenswrapper[4696]: I0321 08:28:49.480288 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:50 crc kubenswrapper[4696]: I0321 08:28:50.482282 4696 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 08:28:51 crc kubenswrapper[4696]: I0321 08:28:51.073892 4696 csr.go:261] certificate signing request csr-8qclp is approved, waiting to be issued Mar 21 08:28:51 crc kubenswrapper[4696]: I0321 08:28:51.083129 4696 csr.go:257] certificate signing request csr-8qclp is issued Mar 21 08:28:51 crc kubenswrapper[4696]: I0321 08:28:51.130851 4696 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 21 08:28:51 crc kubenswrapper[4696]: I0321 08:28:51.215591 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:28:51 crc kubenswrapper[4696]: I0321 08:28:51.215880 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:51 crc kubenswrapper[4696]: I0321 08:28:51.217415 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:51 crc kubenswrapper[4696]: I0321 08:28:51.217482 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:51 crc kubenswrapper[4696]: I0321 08:28:51.217502 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:51 crc kubenswrapper[4696]: I0321 08:28:51.218577 4696 scope.go:117] "RemoveContainer" containerID="a78fedb8a70e8b5cec10a972e63964e900746253f6add19cbedf2e47af762244" Mar 21 08:28:51 crc kubenswrapper[4696]: E0321 08:28:51.218925 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 08:28:51 crc kubenswrapper[4696]: I0321 08:28:51.360903 4696 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 21 08:28:52 crc kubenswrapper[4696]: I0321 08:28:52.084900 4696 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-16 20:08:00.527872679 +0000 UTC Mar 21 08:28:52 crc kubenswrapper[4696]: I0321 08:28:52.084986 4696 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6491h39m8.442897106s for next certificate rotation Mar 21 08:28:52 crc kubenswrapper[4696]: I0321 08:28:52.533800 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:52 crc kubenswrapper[4696]: I0321 08:28:52.534845 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:52 crc kubenswrapper[4696]: I0321 08:28:52.534882 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:52 crc kubenswrapper[4696]: I0321 08:28:52.534896 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.838511 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.840031 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.840080 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.840091 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.840192 4696 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.848164 4696 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.848595 4696 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 21 08:28:55 crc kubenswrapper[4696]: E0321 08:28:55.848640 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.851957 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.851996 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.852006 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.852022 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.852032 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:28:55Z","lastTransitionTime":"2026-03-21T08:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:28:55 crc kubenswrapper[4696]: E0321 08:28:55.864381 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7ca94532-e7e5-44cd-8187-860d6872e1ea\\\",\\\"systemUUID\\\":\\\"fa0c4141-962b-4589-98ba-ecb07b216fa9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.871405 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.871432 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.871439 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.871453 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.871462 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:28:55Z","lastTransitionTime":"2026-03-21T08:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:28:55 crc kubenswrapper[4696]: E0321 08:28:55.881405 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7ca94532-e7e5-44cd-8187-860d6872e1ea\\\",\\\"systemUUID\\\":\\\"fa0c4141-962b-4589-98ba-ecb07b216fa9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.890155 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.890206 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.890217 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.890234 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.890246 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:28:55Z","lastTransitionTime":"2026-03-21T08:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:28:55 crc kubenswrapper[4696]: E0321 08:28:55.899472 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7ca94532-e7e5-44cd-8187-860d6872e1ea\\\",\\\"systemUUID\\\":\\\"fa0c4141-962b-4589-98ba-ecb07b216fa9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.906105 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.906179 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.906203 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.906233 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:28:55 crc kubenswrapper[4696]: I0321 08:28:55.906250 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:28:55Z","lastTransitionTime":"2026-03-21T08:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:28:55 crc kubenswrapper[4696]: E0321 08:28:55.920486 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:28:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7ca94532-e7e5-44cd-8187-860d6872e1ea\\\",\\\"systemUUID\\\":\\\"fa0c4141-962b-4589-98ba-ecb07b216fa9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 08:28:55 crc kubenswrapper[4696]: E0321 08:28:55.920637 4696 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 21 08:28:55 crc kubenswrapper[4696]: E0321 08:28:55.920671 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:56 crc kubenswrapper[4696]: E0321 08:28:56.021245 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:56 crc kubenswrapper[4696]: E0321 08:28:56.122395 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:56 crc kubenswrapper[4696]: E0321 08:28:56.223582 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:56 crc kubenswrapper[4696]: E0321 08:28:56.324415 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:56 crc kubenswrapper[4696]: E0321 08:28:56.424855 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:56 crc kubenswrapper[4696]: E0321 08:28:56.525893 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:56 crc kubenswrapper[4696]: E0321 08:28:56.613110 4696 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 08:28:56 crc kubenswrapper[4696]: E0321 08:28:56.626096 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:56 crc kubenswrapper[4696]: E0321 08:28:56.726905 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:56 crc kubenswrapper[4696]: E0321 08:28:56.827264 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:56 crc kubenswrapper[4696]: E0321 08:28:56.928195 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:57 crc kubenswrapper[4696]: E0321 08:28:57.029194 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:57 crc kubenswrapper[4696]: E0321 08:28:57.130793 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:57 crc kubenswrapper[4696]: E0321 08:28:57.231389 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:57 crc kubenswrapper[4696]: E0321 08:28:57.332177 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:57 crc kubenswrapper[4696]: E0321 08:28:57.432365 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:57 crc kubenswrapper[4696]: E0321 08:28:57.533108 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:57 crc kubenswrapper[4696]: E0321 08:28:57.633909 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:57 crc kubenswrapper[4696]: E0321 08:28:57.735072 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:57 crc kubenswrapper[4696]: E0321 08:28:57.835744 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:57 crc kubenswrapper[4696]: E0321 08:28:57.936476 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:58 crc kubenswrapper[4696]: E0321 08:28:58.036582 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:58 crc kubenswrapper[4696]: E0321 08:28:58.138588 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:58 crc kubenswrapper[4696]: E0321 08:28:58.239546 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:58 crc kubenswrapper[4696]: I0321 08:28:58.250212 4696 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 21 08:28:58 crc kubenswrapper[4696]: E0321 08:28:58.340353 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:58 crc kubenswrapper[4696]: E0321 08:28:58.441373 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:58 crc kubenswrapper[4696]: E0321 08:28:58.542037 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:58 crc kubenswrapper[4696]: E0321 08:28:58.643136 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:58 crc kubenswrapper[4696]: E0321 08:28:58.744045 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:58 crc kubenswrapper[4696]: E0321 08:28:58.844189 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:58 crc kubenswrapper[4696]: E0321 08:28:58.944345 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:59 crc kubenswrapper[4696]: E0321 08:28:59.045404 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:59 crc kubenswrapper[4696]: E0321 08:28:59.145566 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:59 crc kubenswrapper[4696]: E0321 08:28:59.245881 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:59 crc kubenswrapper[4696]: E0321 08:28:59.346341 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:59 crc kubenswrapper[4696]: E0321 08:28:59.446602 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:59 crc kubenswrapper[4696]: E0321 08:28:59.547661 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:59 crc kubenswrapper[4696]: E0321 08:28:59.648034 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:59 crc kubenswrapper[4696]: E0321 08:28:59.749011 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:59 crc kubenswrapper[4696]: E0321 08:28:59.849584 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:28:59 crc kubenswrapper[4696]: E0321 08:28:59.950249 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:00 crc kubenswrapper[4696]: E0321 08:29:00.050995 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:00 crc kubenswrapper[4696]: E0321 08:29:00.151152 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:00 crc kubenswrapper[4696]: E0321 08:29:00.251433 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:00 crc kubenswrapper[4696]: E0321 08:29:00.351555 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:00 crc kubenswrapper[4696]: E0321 08:29:00.451963 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:00 crc kubenswrapper[4696]: E0321 08:29:00.553032 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:00 crc kubenswrapper[4696]: E0321 08:29:00.654064 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:00 crc kubenswrapper[4696]: E0321 08:29:00.754716 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:00 crc kubenswrapper[4696]: E0321 08:29:00.855974 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:00 crc kubenswrapper[4696]: E0321 08:29:00.957046 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:01 crc kubenswrapper[4696]: E0321 08:29:01.057123 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:01 crc kubenswrapper[4696]: E0321 08:29:01.157648 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:01 crc kubenswrapper[4696]: E0321 08:29:01.258523 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:01 crc kubenswrapper[4696]: E0321 08:29:01.358925 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:01 crc kubenswrapper[4696]: E0321 08:29:01.459409 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:01 crc kubenswrapper[4696]: E0321 08:29:01.559572 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:01 crc kubenswrapper[4696]: E0321 08:29:01.659708 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:01 crc kubenswrapper[4696]: E0321 08:29:01.760811 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:01 crc kubenswrapper[4696]: E0321 08:29:01.861599 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:01 crc kubenswrapper[4696]: E0321 08:29:01.962580 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:02 crc kubenswrapper[4696]: E0321 08:29:02.062773 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:02 crc kubenswrapper[4696]: E0321 08:29:02.163778 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:02 crc kubenswrapper[4696]: E0321 08:29:02.264114 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:02 crc kubenswrapper[4696]: E0321 08:29:02.365181 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:02 crc kubenswrapper[4696]: E0321 08:29:02.465732 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:02 crc kubenswrapper[4696]: E0321 08:29:02.566571 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:02 crc kubenswrapper[4696]: E0321 08:29:02.667391 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:02 crc kubenswrapper[4696]: E0321 08:29:02.768709 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:02 crc kubenswrapper[4696]: E0321 08:29:02.869404 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:02 crc kubenswrapper[4696]: E0321 08:29:02.970242 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:03 crc kubenswrapper[4696]: E0321 08:29:03.070739 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:03 crc kubenswrapper[4696]: E0321 08:29:03.171587 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:03 crc kubenswrapper[4696]: E0321 08:29:03.272792 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:03 crc kubenswrapper[4696]: E0321 08:29:03.373812 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:03 crc kubenswrapper[4696]: E0321 08:29:03.474206 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:03 crc kubenswrapper[4696]: I0321 08:29:03.534088 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:29:03 crc kubenswrapper[4696]: I0321 08:29:03.535553 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:03 crc kubenswrapper[4696]: I0321 08:29:03.535601 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:03 crc kubenswrapper[4696]: I0321 08:29:03.535621 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:03 crc kubenswrapper[4696]: I0321 08:29:03.536516 4696 scope.go:117] "RemoveContainer" containerID="a78fedb8a70e8b5cec10a972e63964e900746253f6add19cbedf2e47af762244" Mar 21 08:29:03 crc kubenswrapper[4696]: E0321 08:29:03.536845 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 08:29:03 crc kubenswrapper[4696]: E0321 08:29:03.574381 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:03 crc kubenswrapper[4696]: E0321 08:29:03.674900 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:03 crc kubenswrapper[4696]: E0321 08:29:03.775310 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:03 crc kubenswrapper[4696]: E0321 08:29:03.875545 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:03 crc kubenswrapper[4696]: E0321 08:29:03.976428 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:04 crc kubenswrapper[4696]: E0321 08:29:04.077314 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:04 crc kubenswrapper[4696]: E0321 08:29:04.177552 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:04 crc kubenswrapper[4696]: E0321 08:29:04.278314 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:04 crc kubenswrapper[4696]: E0321 08:29:04.378984 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:04 crc kubenswrapper[4696]: E0321 08:29:04.479588 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:04 crc kubenswrapper[4696]: E0321 08:29:04.579754 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:04 crc kubenswrapper[4696]: E0321 08:29:04.680012 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:04 crc kubenswrapper[4696]: E0321 08:29:04.780290 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:04 crc kubenswrapper[4696]: E0321 08:29:04.880476 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:04 crc kubenswrapper[4696]: E0321 08:29:04.981028 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:05 crc kubenswrapper[4696]: E0321 08:29:05.081394 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:05 crc kubenswrapper[4696]: E0321 08:29:05.182462 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:05 crc kubenswrapper[4696]: E0321 08:29:05.283610 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:05 crc kubenswrapper[4696]: E0321 08:29:05.384167 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:05 crc kubenswrapper[4696]: E0321 08:29:05.485120 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:05 crc kubenswrapper[4696]: E0321 08:29:05.585591 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:05 crc kubenswrapper[4696]: E0321 08:29:05.686305 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:05 crc kubenswrapper[4696]: E0321 08:29:05.787365 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:05 crc kubenswrapper[4696]: E0321 08:29:05.888370 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:05 crc kubenswrapper[4696]: E0321 08:29:05.988703 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:06 crc kubenswrapper[4696]: E0321 08:29:06.088895 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:06 crc kubenswrapper[4696]: E0321 08:29:06.189257 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:06 crc kubenswrapper[4696]: E0321 08:29:06.289791 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:06 crc kubenswrapper[4696]: E0321 08:29:06.290915 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 21 08:29:06 crc kubenswrapper[4696]: I0321 08:29:06.297488 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:06 crc kubenswrapper[4696]: I0321 08:29:06.297560 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:06 crc kubenswrapper[4696]: I0321 08:29:06.297583 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:06 crc kubenswrapper[4696]: I0321 08:29:06.297654 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:06 crc kubenswrapper[4696]: I0321 08:29:06.297676 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:06Z","lastTransitionTime":"2026-03-21T08:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:06 crc kubenswrapper[4696]: E0321 08:29:06.309313 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7ca94532-e7e5-44cd-8187-860d6872e1ea\\\",\\\"systemUUID\\\":\\\"fa0c4141-962b-4589-98ba-ecb07b216fa9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 08:29:06 crc kubenswrapper[4696]: I0321 08:29:06.314589 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:06 crc kubenswrapper[4696]: I0321 08:29:06.314650 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:06 crc kubenswrapper[4696]: I0321 08:29:06.314663 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:06 crc kubenswrapper[4696]: I0321 08:29:06.314685 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:06 crc kubenswrapper[4696]: I0321 08:29:06.314699 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:06Z","lastTransitionTime":"2026-03-21T08:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:06 crc kubenswrapper[4696]: E0321 08:29:06.331130 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7ca94532-e7e5-44cd-8187-860d6872e1ea\\\",\\\"systemUUID\\\":\\\"fa0c4141-962b-4589-98ba-ecb07b216fa9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 08:29:06 crc kubenswrapper[4696]: I0321 08:29:06.335980 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:06 crc kubenswrapper[4696]: I0321 08:29:06.336109 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:06 crc kubenswrapper[4696]: I0321 08:29:06.336128 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:06 crc kubenswrapper[4696]: I0321 08:29:06.336194 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:06 crc kubenswrapper[4696]: I0321 08:29:06.336214 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:06Z","lastTransitionTime":"2026-03-21T08:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:06 crc kubenswrapper[4696]: E0321 08:29:06.353734 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7ca94532-e7e5-44cd-8187-860d6872e1ea\\\",\\\"systemUUID\\\":\\\"fa0c4141-962b-4589-98ba-ecb07b216fa9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 08:29:06 crc kubenswrapper[4696]: I0321 08:29:06.359935 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:06 crc kubenswrapper[4696]: I0321 08:29:06.359986 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:06 crc kubenswrapper[4696]: I0321 08:29:06.360006 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:06 crc kubenswrapper[4696]: I0321 08:29:06.360031 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:06 crc kubenswrapper[4696]: I0321 08:29:06.360052 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:06Z","lastTransitionTime":"2026-03-21T08:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:06 crc kubenswrapper[4696]: E0321 08:29:06.372028 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7ca94532-e7e5-44cd-8187-860d6872e1ea\\\",\\\"systemUUID\\\":\\\"fa0c4141-962b-4589-98ba-ecb07b216fa9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 08:29:06 crc kubenswrapper[4696]: E0321 08:29:06.372412 4696 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 21 08:29:06 crc kubenswrapper[4696]: E0321 08:29:06.390313 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:06 crc kubenswrapper[4696]: E0321 08:29:06.490645 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:06 crc kubenswrapper[4696]: E0321 08:29:06.591292 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:06 crc kubenswrapper[4696]: E0321 08:29:06.614123 4696 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 08:29:06 crc kubenswrapper[4696]: E0321 08:29:06.692000 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:06 crc kubenswrapper[4696]: E0321 08:29:06.793060 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:06 crc kubenswrapper[4696]: E0321 08:29:06.894144 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:06 crc kubenswrapper[4696]: E0321 08:29:06.994718 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:07 crc kubenswrapper[4696]: E0321 08:29:07.095065 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:07 crc kubenswrapper[4696]: E0321 08:29:07.196319 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:07 crc kubenswrapper[4696]: E0321 08:29:07.297497 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:07 crc kubenswrapper[4696]: E0321 08:29:07.397861 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:07 crc kubenswrapper[4696]: E0321 08:29:07.498596 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:07 crc kubenswrapper[4696]: E0321 08:29:07.599738 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:07 crc kubenswrapper[4696]: E0321 08:29:07.699945 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:07 crc kubenswrapper[4696]: E0321 08:29:07.800634 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:07 crc kubenswrapper[4696]: E0321 08:29:07.900878 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:08 crc kubenswrapper[4696]: E0321 08:29:08.001885 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:08 crc kubenswrapper[4696]: E0321 08:29:08.103027 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:08 crc kubenswrapper[4696]: E0321 08:29:08.204105 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:08 crc kubenswrapper[4696]: E0321 08:29:08.304808 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:08 crc kubenswrapper[4696]: E0321 08:29:08.405223 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:08 crc kubenswrapper[4696]: E0321 08:29:08.506186 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:08 crc kubenswrapper[4696]: I0321 08:29:08.534337 4696 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 21 08:29:08 crc kubenswrapper[4696]: I0321 08:29:08.536235 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:08 crc kubenswrapper[4696]: I0321 08:29:08.536343 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:08 crc kubenswrapper[4696]: I0321 08:29:08.536370 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:08 crc kubenswrapper[4696]: E0321 08:29:08.607281 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:08 crc kubenswrapper[4696]: E0321 08:29:08.707767 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:08 crc kubenswrapper[4696]: E0321 08:29:08.809274 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:08 crc kubenswrapper[4696]: E0321 08:29:08.910452 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:09 crc kubenswrapper[4696]: E0321 08:29:09.010615 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:09 crc kubenswrapper[4696]: E0321 08:29:09.111478 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:09 crc kubenswrapper[4696]: E0321 08:29:09.212485 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:09 crc kubenswrapper[4696]: E0321 08:29:09.313416 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:09 crc kubenswrapper[4696]: E0321 08:29:09.414541 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:09 crc kubenswrapper[4696]: E0321 08:29:09.515690 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:09 crc kubenswrapper[4696]: E0321 08:29:09.615932 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:09 crc kubenswrapper[4696]: E0321 08:29:09.716726 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:09 crc kubenswrapper[4696]: E0321 08:29:09.817539 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:09 crc kubenswrapper[4696]: E0321 08:29:09.917989 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:10 crc kubenswrapper[4696]: E0321 08:29:10.018784 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:10 crc kubenswrapper[4696]: E0321 08:29:10.119418 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:10 crc kubenswrapper[4696]: E0321 08:29:10.220704 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:10 crc kubenswrapper[4696]: E0321 08:29:10.321659 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:10 crc kubenswrapper[4696]: E0321 08:29:10.422058 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:10 crc kubenswrapper[4696]: E0321 08:29:10.522435 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:10 crc kubenswrapper[4696]: E0321 08:29:10.623287 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:10 crc kubenswrapper[4696]: E0321 08:29:10.724167 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:10 crc kubenswrapper[4696]: I0321 08:29:10.730614 4696 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 21 08:29:10 crc kubenswrapper[4696]: E0321 08:29:10.824857 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:10 crc kubenswrapper[4696]: E0321 08:29:10.925247 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:11 crc kubenswrapper[4696]: E0321 08:29:11.026415 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:11 crc kubenswrapper[4696]: E0321 08:29:11.127329 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:11 crc kubenswrapper[4696]: E0321 08:29:11.227448 4696 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.256487 4696 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.330023 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.330088 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.330105 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.330122 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.330135 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:11Z","lastTransitionTime":"2026-03-21T08:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.433654 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.433712 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.433732 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.433762 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.433785 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:11Z","lastTransitionTime":"2026-03-21T08:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.505789 4696 apiserver.go:52] "Watching apiserver" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.511967 4696 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.512487 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.512994 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.513266 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:11 crc kubenswrapper[4696]: E0321 08:29:11.513481 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.514209 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.514396 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:11 crc kubenswrapper[4696]: E0321 08:29:11.514482 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.514598 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.513987 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:11 crc kubenswrapper[4696]: E0321 08:29:11.515314 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.516519 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.517344 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.517480 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.517518 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.517559 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.517652 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.518235 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.518281 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.520034 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.536543 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.537046 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.537072 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.537145 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.537171 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:11Z","lastTransitionTime":"2026-03-21T08:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.559144 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.570524 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.580373 4696 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.583382 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.593506 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.604385 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.613747 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.621767 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.639486 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.639713 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.639812 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.639941 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.640040 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:11Z","lastTransitionTime":"2026-03-21T08:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669352 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669396 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669415 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669431 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669447 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669464 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669479 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669499 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669517 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669532 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669548 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669565 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669581 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669604 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669620 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669636 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669652 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669669 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669684 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669702 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669718 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669734 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669753 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669781 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669803 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669849 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669869 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669884 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669900 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669923 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669938 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669955 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669971 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.669986 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670011 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670001 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670033 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670127 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670164 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670197 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670234 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670267 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670301 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670334 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670367 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670399 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670431 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670466 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670498 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670530 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670562 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670592 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670628 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670660 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670691 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670724 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670755 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670788 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670851 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670996 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671031 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671065 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671099 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671131 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671162 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671196 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671227 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671258 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671291 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671324 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671363 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671396 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671429 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671460 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671492 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671523 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671558 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671592 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671626 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671658 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671691 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671730 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671762 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671795 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671851 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672006 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672037 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672071 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672105 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672141 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672173 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672206 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672238 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672272 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672305 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672337 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672369 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672405 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672437 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672472 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672507 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672539 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672570 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672604 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672640 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672675 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672708 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672742 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672778 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672813 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672871 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672903 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672938 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672976 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673008 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673042 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673074 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673109 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673141 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673175 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673211 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673244 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673282 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673317 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673350 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673383 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673416 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673451 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673484 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673522 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673558 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673592 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673626 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673662 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673697 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673731 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673765 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673799 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673860 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673894 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673928 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673963 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673997 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674031 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674065 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674100 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674132 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674166 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674199 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674233 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674265 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674299 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674331 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674366 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674399 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674432 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674467 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674500 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674535 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674570 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674605 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674641 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674674 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674708 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674743 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674778 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674813 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675054 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675088 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675122 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675161 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675195 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675231 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675266 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675302 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675344 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675386 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675428 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675464 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675499 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675533 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675569 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675607 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675641 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675675 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675710 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675743 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675780 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675840 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675874 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675912 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675969 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676003 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676040 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676074 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676108 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676171 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676221 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676263 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676300 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676341 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676379 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676417 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676453 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676491 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676530 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676567 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676605 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676642 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676683 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676758 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.678341 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670462 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.679574 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.679623 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670663 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670701 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670891 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671070 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671082 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671148 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671332 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.680314 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.681096 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.681164 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.681124 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.681241 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.681662 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.682135 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.682467 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.682699 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: E0321 08:29:11.682755 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:12.18272915 +0000 UTC m=+86.303609863 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.683064 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.683202 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.683236 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.683288 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.683426 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.683653 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.684188 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.684194 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671489 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671623 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671635 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671657 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671793 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671901 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.684419 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671908 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671923 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672021 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672193 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672214 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672464 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672539 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672635 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672645 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672708 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672785 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672802 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.672953 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673039 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673059 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673166 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673287 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673287 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673385 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673491 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.673679 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674455 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.674849 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675184 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675283 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675442 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.675745 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676371 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676448 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676615 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676750 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.676988 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.677082 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.677275 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.677590 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.677783 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.677951 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.678165 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.678192 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.678244 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.678257 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.678286 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.678766 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.678767 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.679119 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.679249 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.679264 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.679304 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.679661 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.679912 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.680008 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.680094 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.680225 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.680269 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.684684 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.684915 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.685061 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.685246 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.685548 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.685870 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.685890 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.685671 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.686514 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.670513 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.688737 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.689283 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.689154 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.689344 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.689359 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.689368 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.689377 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.689382 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.689393 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.689412 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.689494 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.689802 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.689914 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.690013 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.690236 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.690297 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.690410 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.690948 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.691291 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.691430 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.690639 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.692254 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.692260 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.692331 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.692349 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.692485 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.692645 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.692770 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.692868 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.692895 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.692938 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: E0321 08:29:11.692982 4696 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.693028 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.693269 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: E0321 08:29:11.693092 4696 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.693276 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.693401 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: E0321 08:29:11.693428 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:12.193406767 +0000 UTC m=+86.314287480 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.693753 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.693876 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.693886 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.694013 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.694146 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.694225 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.694258 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.694372 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.694387 4696 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.694586 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.694608 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.694612 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.694950 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.694968 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 21 08:29:11 crc kubenswrapper[4696]: E0321 08:29:11.695081 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:12.19505262 +0000 UTC m=+86.315933373 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.695122 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.695182 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.695271 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.695413 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.695424 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.695523 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.695895 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.696000 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.696292 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.696344 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.671309 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.696549 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.696770 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.696863 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.697114 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.697533 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.697706 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.697855 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.697958 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.698039 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.698068 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.702792 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 21 08:29:11 crc kubenswrapper[4696]: E0321 08:29:11.709571 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 08:29:11 crc kubenswrapper[4696]: E0321 08:29:11.709606 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 08:29:11 crc kubenswrapper[4696]: E0321 08:29:11.709625 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 08:29:11 crc kubenswrapper[4696]: E0321 08:29:11.709640 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 08:29:11 crc kubenswrapper[4696]: E0321 08:29:11.709652 4696 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:11 crc kubenswrapper[4696]: E0321 08:29:11.709659 4696 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:11 crc kubenswrapper[4696]: E0321 08:29:11.709730 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:12.209709902 +0000 UTC m=+86.330590635 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:11 crc kubenswrapper[4696]: E0321 08:29:11.709756 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:12.209745912 +0000 UTC m=+86.330626645 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.709878 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.710288 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.710395 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.710649 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.710741 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.711080 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.711125 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.712005 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.718864 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.719247 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.720134 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.720556 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.720953 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.721008 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.722421 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.722570 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.722812 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.722879 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.723323 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.723655 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.723667 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.724140 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.724189 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.726058 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.727507 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.728158 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.731655 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.734455 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.735071 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.742308 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.743025 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.743062 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.743073 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.743090 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.743102 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:11Z","lastTransitionTime":"2026-03-21T08:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.744421 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.745671 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778082 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778403 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778550 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778562 4696 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778632 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778645 4696 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778656 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778666 4696 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778676 4696 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778688 4696 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778700 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778710 4696 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778719 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778256 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778728 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778789 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778802 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778812 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778839 4696 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778850 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778860 4696 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778870 4696 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778879 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778887 4696 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778895 4696 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778903 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778912 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778920 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778928 4696 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778936 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778944 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778952 4696 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778960 4696 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778968 4696 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778976 4696 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778984 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778992 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.778999 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779008 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779015 4696 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779023 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779031 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779041 4696 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779064 4696 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779072 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779082 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779092 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779100 4696 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779111 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779119 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779129 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779137 4696 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779145 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779153 4696 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779161 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779170 4696 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779178 4696 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779187 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779196 4696 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779205 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779215 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779235 4696 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779244 4696 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779253 4696 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779262 4696 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779270 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779279 4696 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779287 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779296 4696 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779304 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779313 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779321 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779329 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779338 4696 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779347 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779355 4696 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779363 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779372 4696 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779381 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779389 4696 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779397 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779405 4696 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779413 4696 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779422 4696 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779431 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779439 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779449 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779459 4696 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779467 4696 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779475 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779484 4696 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779492 4696 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779501 4696 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779509 4696 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779517 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779526 4696 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779533 4696 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779541 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779549 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779557 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779565 4696 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779574 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779582 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779590 4696 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779598 4696 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779607 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779615 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779623 4696 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779631 4696 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779640 4696 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779647 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779655 4696 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779663 4696 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779671 4696 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779680 4696 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779689 4696 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779698 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779706 4696 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779715 4696 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779723 4696 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779730 4696 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779738 4696 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779746 4696 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779754 4696 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779762 4696 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779771 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779780 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779788 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779796 4696 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779803 4696 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779827 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779835 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779843 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779851 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779858 4696 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779866 4696 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779874 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779881 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779889 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779898 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779907 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779914 4696 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779922 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779935 4696 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779944 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779952 4696 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779960 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779968 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779977 4696 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779985 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.779993 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780000 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780008 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780016 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780024 4696 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780032 4696 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780040 4696 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780048 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780056 4696 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780064 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780072 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780080 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780091 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780098 4696 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780106 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780115 4696 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780124 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780133 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780141 4696 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780149 4696 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780157 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780165 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780173 4696 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780181 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780189 4696 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780197 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780205 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780212 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780221 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780229 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780237 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780246 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780255 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780263 4696 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780271 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780279 4696 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780287 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780295 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780304 4696 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780313 4696 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780321 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780330 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780338 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780347 4696 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780357 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780365 4696 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.780373 4696 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.835339 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.845096 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.845157 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.845180 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.845204 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.845272 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:11Z","lastTransitionTime":"2026-03-21T08:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.845373 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.851372 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 21 08:29:11 crc kubenswrapper[4696]: W0321 08:29:11.865860 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-6c671fe2655b9a7035dc4e499857d236c574047d16f98e2631991cc343afdea4 WatchSource:0}: Error finding container 6c671fe2655b9a7035dc4e499857d236c574047d16f98e2631991cc343afdea4: Status 404 returned error can't find the container with id 6c671fe2655b9a7035dc4e499857d236c574047d16f98e2631991cc343afdea4 Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.910256 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"32adecaa9b31a4d3acc7bf9c57faf27f1707fe2d19ca7123fbd63e47f37c11c3"} Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.913105 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6c671fe2655b9a7035dc4e499857d236c574047d16f98e2631991cc343afdea4"} Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.913997 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"6ae89a887b97a20a5ef7cd8ed745f5907e2840b2ce9272dff80369e2e704e7c7"} Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.948481 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.948524 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.948538 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.948556 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:11 crc kubenswrapper[4696]: I0321 08:29:11.948567 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:11Z","lastTransitionTime":"2026-03-21T08:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.051448 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.051791 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.051800 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.051834 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.051845 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:12Z","lastTransitionTime":"2026-03-21T08:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.153932 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.153969 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.153979 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.153994 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.154005 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:12Z","lastTransitionTime":"2026-03-21T08:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.184505 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:12 crc kubenswrapper[4696]: E0321 08:29:12.184734 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:13.184700908 +0000 UTC m=+87.305581621 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.256573 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.256637 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.256654 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.256676 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.256695 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:12Z","lastTransitionTime":"2026-03-21T08:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.285878 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.285929 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.285949 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.285972 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:12 crc kubenswrapper[4696]: E0321 08:29:12.286036 4696 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 08:29:12 crc kubenswrapper[4696]: E0321 08:29:12.286088 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:13.286075775 +0000 UTC m=+87.406956488 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 08:29:12 crc kubenswrapper[4696]: E0321 08:29:12.286203 4696 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 08:29:12 crc kubenswrapper[4696]: E0321 08:29:12.286273 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 08:29:12 crc kubenswrapper[4696]: E0321 08:29:12.286316 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 08:29:12 crc kubenswrapper[4696]: E0321 08:29:12.286329 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:13.286303471 +0000 UTC m=+87.407184214 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 08:29:12 crc kubenswrapper[4696]: E0321 08:29:12.286338 4696 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:12 crc kubenswrapper[4696]: E0321 08:29:12.286410 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:13.286378773 +0000 UTC m=+87.407259516 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:12 crc kubenswrapper[4696]: E0321 08:29:12.286520 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 08:29:12 crc kubenswrapper[4696]: E0321 08:29:12.286552 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 08:29:12 crc kubenswrapper[4696]: E0321 08:29:12.286568 4696 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:12 crc kubenswrapper[4696]: E0321 08:29:12.286622 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:13.286607889 +0000 UTC m=+87.407488642 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.359261 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.359321 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.359332 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.359348 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.359358 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:12Z","lastTransitionTime":"2026-03-21T08:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.462312 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.462381 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.462395 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.462418 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.462442 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:12Z","lastTransitionTime":"2026-03-21T08:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.540629 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.543476 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.545667 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.548345 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.549755 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.551967 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.553555 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.555268 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.557788 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.561234 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.562460 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.565115 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.565193 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.565217 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.565245 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.565268 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:12Z","lastTransitionTime":"2026-03-21T08:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.565555 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.567736 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.571071 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.572384 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.574367 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.576214 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.577180 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.579191 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.580538 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.581626 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.583028 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.583663 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.585759 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.586674 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.588138 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.589052 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.589795 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.591322 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.592185 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.594167 4696 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.594361 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.597535 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.599969 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.600590 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.603562 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.605385 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.606203 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.607795 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.608788 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.609473 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.610995 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.612327 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.613175 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.614394 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.615146 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.616513 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.617631 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.618574 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.619090 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.619516 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.620388 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.620993 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.621839 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.667868 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.667910 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.667921 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.667935 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.667946 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:12Z","lastTransitionTime":"2026-03-21T08:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.770705 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.770730 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.770746 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.770760 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.770770 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:12Z","lastTransitionTime":"2026-03-21T08:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.873393 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.873474 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.873488 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.873511 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.873523 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:12Z","lastTransitionTime":"2026-03-21T08:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.918303 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"758bc9ea912dc40109390cae7de1fc162220b8a3c22d08d13aa66dd5572b8580"} Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.918346 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c68ef4964aba47ed6e3a72cd542a1701c44bfb966830bf98ad0ea958c6e425c1"} Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.920039 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"2c169ec3b2de6b14acbf2670a5b699af60b44e866a4112d436d785029b8ee1eb"} Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.935189 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:29:12Z is after 2025-08-24T17:21:41Z" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.947603 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:29:12Z is after 2025-08-24T17:21:41Z" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.962206 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:29:12Z is after 2025-08-24T17:21:41Z" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.976045 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.976084 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.976096 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.976137 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.976149 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:12Z","lastTransitionTime":"2026-03-21T08:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.979909 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://758bc9ea912dc40109390cae7de1fc162220b8a3c22d08d13aa66dd5572b8580\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T08:29:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c68ef4964aba47ed6e3a72cd542a1701c44bfb966830bf98ad0ea958c6e425c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T08:29:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:29:12Z is after 2025-08-24T17:21:41Z" Mar 21 08:29:12 crc kubenswrapper[4696]: I0321 08:29:12.994851 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:29:12Z is after 2025-08-24T17:21:41Z" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.007343 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:29:13Z is after 2025-08-24T17:21:41Z" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.020746 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://758bc9ea912dc40109390cae7de1fc162220b8a3c22d08d13aa66dd5572b8580\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T08:29:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c68ef4964aba47ed6e3a72cd542a1701c44bfb966830bf98ad0ea958c6e425c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T08:29:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:29:13Z is after 2025-08-24T17:21:41Z" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.038452 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:29:13Z is after 2025-08-24T17:21:41Z" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.054512 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:29:13Z is after 2025-08-24T17:21:41Z" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.067168 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c169ec3b2de6b14acbf2670a5b699af60b44e866a4112d436d785029b8ee1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T08:29:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:29:13Z is after 2025-08-24T17:21:41Z" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.078510 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:29:13Z is after 2025-08-24T17:21:41Z" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.078718 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.078733 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.078741 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.078755 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.078764 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:13Z","lastTransitionTime":"2026-03-21T08:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.093013 4696 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T08:29:11Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-21T08:29:13Z is after 2025-08-24T17:21:41Z" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.181040 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.181071 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.181079 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.181091 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.181101 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:13Z","lastTransitionTime":"2026-03-21T08:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.193703 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:13 crc kubenswrapper[4696]: E0321 08:29:13.193898 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:15.193869261 +0000 UTC m=+89.314749974 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.284243 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.284282 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.284292 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.284309 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.284323 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:13Z","lastTransitionTime":"2026-03-21T08:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.294949 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.294989 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.295020 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.295063 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:13 crc kubenswrapper[4696]: E0321 08:29:13.295166 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 08:29:13 crc kubenswrapper[4696]: E0321 08:29:13.295181 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 08:29:13 crc kubenswrapper[4696]: E0321 08:29:13.295192 4696 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:13 crc kubenswrapper[4696]: E0321 08:29:13.295233 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:15.295221807 +0000 UTC m=+89.416102510 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:13 crc kubenswrapper[4696]: E0321 08:29:13.295524 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 08:29:13 crc kubenswrapper[4696]: E0321 08:29:13.295535 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 08:29:13 crc kubenswrapper[4696]: E0321 08:29:13.295544 4696 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:13 crc kubenswrapper[4696]: E0321 08:29:13.295568 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:15.295559397 +0000 UTC m=+89.416440110 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:13 crc kubenswrapper[4696]: E0321 08:29:13.295606 4696 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 08:29:13 crc kubenswrapper[4696]: E0321 08:29:13.295628 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:15.295619108 +0000 UTC m=+89.416499821 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 08:29:13 crc kubenswrapper[4696]: E0321 08:29:13.295650 4696 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 08:29:13 crc kubenswrapper[4696]: E0321 08:29:13.295668 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:15.295663179 +0000 UTC m=+89.416543892 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.386402 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.386448 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.386457 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.386470 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.386479 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:13Z","lastTransitionTime":"2026-03-21T08:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.488234 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.488265 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.488273 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.488284 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.488293 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:13Z","lastTransitionTime":"2026-03-21T08:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.533569 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:13 crc kubenswrapper[4696]: E0321 08:29:13.533714 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.534036 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:13 crc kubenswrapper[4696]: E0321 08:29:13.534084 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.534121 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:13 crc kubenswrapper[4696]: E0321 08:29:13.534160 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.590644 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.590672 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.590680 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.590692 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.590701 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:13Z","lastTransitionTime":"2026-03-21T08:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.693472 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.693502 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.693510 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.693523 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.693532 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:13Z","lastTransitionTime":"2026-03-21T08:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.795281 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.795315 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.795324 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.795337 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.795348 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:13Z","lastTransitionTime":"2026-03-21T08:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.805921 4696 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.897741 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.897780 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.897791 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.897807 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:13 crc kubenswrapper[4696]: I0321 08:29:13.897837 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:13Z","lastTransitionTime":"2026-03-21T08:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.000195 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.000230 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.000241 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.000256 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.000268 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:14Z","lastTransitionTime":"2026-03-21T08:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.102559 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.102598 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.102609 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.102623 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.102632 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:14Z","lastTransitionTime":"2026-03-21T08:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.205435 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.205509 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.205530 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.205548 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.205561 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:14Z","lastTransitionTime":"2026-03-21T08:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.307850 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.307924 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.307939 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.307961 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.307984 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:14Z","lastTransitionTime":"2026-03-21T08:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.410544 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.410642 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.410659 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.410688 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.410709 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:14Z","lastTransitionTime":"2026-03-21T08:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.513973 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.514027 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.514041 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.514060 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.514073 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:14Z","lastTransitionTime":"2026-03-21T08:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.617472 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.617522 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.617531 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.617547 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.617557 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:14Z","lastTransitionTime":"2026-03-21T08:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.720639 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.720707 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.720718 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.720735 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.720745 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:14Z","lastTransitionTime":"2026-03-21T08:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.824360 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.824433 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.824447 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.824470 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.824482 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:14Z","lastTransitionTime":"2026-03-21T08:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.926719 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.926770 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.926780 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.926802 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.926873 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:14Z","lastTransitionTime":"2026-03-21T08:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:14 crc kubenswrapper[4696]: I0321 08:29:14.930017 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"4007abc24c91b9c0e2a6c49c1cf71bc98c66d0f8ebaf6aec08790bbcdc0aef04"} Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.029448 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.029493 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.029507 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.029524 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.029549 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:15Z","lastTransitionTime":"2026-03-21T08:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.132612 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.132639 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.132647 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.132659 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.132667 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:15Z","lastTransitionTime":"2026-03-21T08:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.213685 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:15 crc kubenswrapper[4696]: E0321 08:29:15.213790 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:19.213765188 +0000 UTC m=+93.334645901 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.235950 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.235997 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.236008 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.236021 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.236032 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:15Z","lastTransitionTime":"2026-03-21T08:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.315087 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.315146 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.315172 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.315198 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:15 crc kubenswrapper[4696]: E0321 08:29:15.315284 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 08:29:15 crc kubenswrapper[4696]: E0321 08:29:15.315310 4696 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 08:29:15 crc kubenswrapper[4696]: E0321 08:29:15.315341 4696 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 08:29:15 crc kubenswrapper[4696]: E0321 08:29:15.315374 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:19.315355161 +0000 UTC m=+93.436235874 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 08:29:15 crc kubenswrapper[4696]: E0321 08:29:15.315319 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 08:29:15 crc kubenswrapper[4696]: E0321 08:29:15.315401 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:19.315385122 +0000 UTC m=+93.436265835 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 08:29:15 crc kubenswrapper[4696]: E0321 08:29:15.315284 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 08:29:15 crc kubenswrapper[4696]: E0321 08:29:15.315414 4696 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:15 crc kubenswrapper[4696]: E0321 08:29:15.315423 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 08:29:15 crc kubenswrapper[4696]: E0321 08:29:15.315431 4696 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:15 crc kubenswrapper[4696]: E0321 08:29:15.315449 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:19.315439253 +0000 UTC m=+93.436319966 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:15 crc kubenswrapper[4696]: E0321 08:29:15.315462 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:19.315456603 +0000 UTC m=+93.436337316 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.338903 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.338942 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.338951 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.338966 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.338976 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:15Z","lastTransitionTime":"2026-03-21T08:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.442245 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.442309 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.442322 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.442344 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.442359 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:15Z","lastTransitionTime":"2026-03-21T08:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.533739 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.533782 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:15 crc kubenswrapper[4696]: E0321 08:29:15.533944 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.533974 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:15 crc kubenswrapper[4696]: E0321 08:29:15.534169 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 08:29:15 crc kubenswrapper[4696]: E0321 08:29:15.534404 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.544851 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.544895 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.544904 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.544926 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.544940 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:15Z","lastTransitionTime":"2026-03-21T08:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.648183 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.648498 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.648676 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.648854 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.649004 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:15Z","lastTransitionTime":"2026-03-21T08:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.750734 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.750777 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.750788 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.750805 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.750934 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:15Z","lastTransitionTime":"2026-03-21T08:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.853054 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.853150 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.853169 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.853196 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.853214 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:15Z","lastTransitionTime":"2026-03-21T08:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.956057 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.956101 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.956111 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.956125 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:15 crc kubenswrapper[4696]: I0321 08:29:15.956135 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:15Z","lastTransitionTime":"2026-03-21T08:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.059680 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.059748 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.059869 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.059900 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.059920 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:16Z","lastTransitionTime":"2026-03-21T08:29:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.162545 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.162588 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.162599 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.162614 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.162623 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:16Z","lastTransitionTime":"2026-03-21T08:29:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.266419 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.266487 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.266498 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.266514 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.266524 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:16Z","lastTransitionTime":"2026-03-21T08:29:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.369163 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.369446 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.369457 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.369472 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.369482 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:16Z","lastTransitionTime":"2026-03-21T08:29:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.471620 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.471647 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.471655 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.471668 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.471676 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:16Z","lastTransitionTime":"2026-03-21T08:29:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.573856 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.573900 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.573913 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.573929 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.573938 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:16Z","lastTransitionTime":"2026-03-21T08:29:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.581849 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.581914 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.581932 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.581956 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 08:29:16 crc kubenswrapper[4696]: I0321 08:29:16.581979 4696 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T08:29:16Z","lastTransitionTime":"2026-03-21T08:29:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 08:29:17 crc kubenswrapper[4696]: I0321 08:29:17.531615 4696 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 21 08:29:17 crc kubenswrapper[4696]: I0321 08:29:17.534009 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:17 crc kubenswrapper[4696]: I0321 08:29:17.534039 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:17 crc kubenswrapper[4696]: E0321 08:29:17.534142 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 08:29:17 crc kubenswrapper[4696]: I0321 08:29:17.534023 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:17 crc kubenswrapper[4696]: E0321 08:29:17.534564 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 08:29:17 crc kubenswrapper[4696]: E0321 08:29:17.534610 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 08:29:17 crc kubenswrapper[4696]: I0321 08:29:17.542118 4696 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 21 08:29:18 crc kubenswrapper[4696]: I0321 08:29:18.554005 4696 scope.go:117] "RemoveContainer" containerID="a78fedb8a70e8b5cec10a972e63964e900746253f6add19cbedf2e47af762244" Mar 21 08:29:18 crc kubenswrapper[4696]: I0321 08:29:18.554045 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 21 08:29:18 crc kubenswrapper[4696]: E0321 08:29:18.554189 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 08:29:18 crc kubenswrapper[4696]: I0321 08:29:18.941212 4696 scope.go:117] "RemoveContainer" containerID="a78fedb8a70e8b5cec10a972e63964e900746253f6add19cbedf2e47af762244" Mar 21 08:29:18 crc kubenswrapper[4696]: E0321 08:29:18.941334 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 21 08:29:19 crc kubenswrapper[4696]: I0321 08:29:19.248165 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:19 crc kubenswrapper[4696]: E0321 08:29:19.248273 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:27.248250423 +0000 UTC m=+101.369131136 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:19 crc kubenswrapper[4696]: I0321 08:29:19.349489 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:19 crc kubenswrapper[4696]: I0321 08:29:19.349524 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:19 crc kubenswrapper[4696]: I0321 08:29:19.349546 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:19 crc kubenswrapper[4696]: I0321 08:29:19.349565 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:19 crc kubenswrapper[4696]: E0321 08:29:19.349627 4696 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 08:29:19 crc kubenswrapper[4696]: E0321 08:29:19.349669 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:27.349657 +0000 UTC m=+101.470537713 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 08:29:19 crc kubenswrapper[4696]: E0321 08:29:19.349670 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 08:29:19 crc kubenswrapper[4696]: E0321 08:29:19.349683 4696 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 08:29:19 crc kubenswrapper[4696]: E0321 08:29:19.349721 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 08:29:19 crc kubenswrapper[4696]: E0321 08:29:19.349772 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 08:29:19 crc kubenswrapper[4696]: E0321 08:29:19.349790 4696 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:19 crc kubenswrapper[4696]: E0321 08:29:19.349732 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:27.349720402 +0000 UTC m=+101.470601115 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 08:29:19 crc kubenswrapper[4696]: E0321 08:29:19.349688 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 08:29:19 crc kubenswrapper[4696]: E0321 08:29:19.349901 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:27.349883637 +0000 UTC m=+101.470764350 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:19 crc kubenswrapper[4696]: E0321 08:29:19.349924 4696 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:19 crc kubenswrapper[4696]: E0321 08:29:19.349989 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:27.34997658 +0000 UTC m=+101.470857293 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:19 crc kubenswrapper[4696]: I0321 08:29:19.533789 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:19 crc kubenswrapper[4696]: I0321 08:29:19.533845 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:19 crc kubenswrapper[4696]: I0321 08:29:19.533863 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:19 crc kubenswrapper[4696]: E0321 08:29:19.533943 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 08:29:19 crc kubenswrapper[4696]: E0321 08:29:19.534006 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 08:29:19 crc kubenswrapper[4696]: E0321 08:29:19.534108 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 08:29:21 crc kubenswrapper[4696]: I0321 08:29:21.534463 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:21 crc kubenswrapper[4696]: I0321 08:29:21.534527 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:21 crc kubenswrapper[4696]: I0321 08:29:21.534498 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:21 crc kubenswrapper[4696]: E0321 08:29:21.534652 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 08:29:21 crc kubenswrapper[4696]: E0321 08:29:21.534919 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 08:29:21 crc kubenswrapper[4696]: E0321 08:29:21.534993 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 08:29:23 crc kubenswrapper[4696]: I0321 08:29:23.534448 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:23 crc kubenswrapper[4696]: E0321 08:29:23.534585 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 08:29:23 crc kubenswrapper[4696]: I0321 08:29:23.534924 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:23 crc kubenswrapper[4696]: E0321 08:29:23.534976 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 08:29:23 crc kubenswrapper[4696]: I0321 08:29:23.535019 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:23 crc kubenswrapper[4696]: E0321 08:29:23.535059 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 08:29:25 crc kubenswrapper[4696]: I0321 08:29:25.534184 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:25 crc kubenswrapper[4696]: I0321 08:29:25.534284 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:25 crc kubenswrapper[4696]: I0321 08:29:25.534199 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:25 crc kubenswrapper[4696]: E0321 08:29:25.534386 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 08:29:25 crc kubenswrapper[4696]: E0321 08:29:25.534566 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 08:29:25 crc kubenswrapper[4696]: E0321 08:29:25.534694 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 08:29:27 crc kubenswrapper[4696]: I0321 08:29:27.324596 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:27 crc kubenswrapper[4696]: E0321 08:29:27.324743 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:43.324718048 +0000 UTC m=+117.445598751 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:27 crc kubenswrapper[4696]: I0321 08:29:27.426720 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:27 crc kubenswrapper[4696]: I0321 08:29:27.426765 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:27 crc kubenswrapper[4696]: I0321 08:29:27.426784 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:27 crc kubenswrapper[4696]: I0321 08:29:27.426803 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:27 crc kubenswrapper[4696]: E0321 08:29:27.426908 4696 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 08:29:27 crc kubenswrapper[4696]: E0321 08:29:27.426965 4696 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 08:29:27 crc kubenswrapper[4696]: E0321 08:29:27.427005 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:43.426986708 +0000 UTC m=+117.547867411 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 08:29:27 crc kubenswrapper[4696]: E0321 08:29:27.427024 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:43.427016389 +0000 UTC m=+117.547897102 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 08:29:27 crc kubenswrapper[4696]: E0321 08:29:27.427032 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 08:29:27 crc kubenswrapper[4696]: E0321 08:29:27.427055 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 08:29:27 crc kubenswrapper[4696]: E0321 08:29:27.427057 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 08:29:27 crc kubenswrapper[4696]: E0321 08:29:27.427068 4696 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:27 crc kubenswrapper[4696]: E0321 08:29:27.427073 4696 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 08:29:27 crc kubenswrapper[4696]: E0321 08:29:27.427083 4696 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:27 crc kubenswrapper[4696]: E0321 08:29:27.427111 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:43.427094071 +0000 UTC m=+117.547974884 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:27 crc kubenswrapper[4696]: E0321 08:29:27.427128 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:43.427120752 +0000 UTC m=+117.548001595 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 08:29:27 crc kubenswrapper[4696]: I0321 08:29:27.534278 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:27 crc kubenswrapper[4696]: E0321 08:29:27.534415 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 08:29:27 crc kubenswrapper[4696]: I0321 08:29:27.534299 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:27 crc kubenswrapper[4696]: E0321 08:29:27.534497 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 08:29:27 crc kubenswrapper[4696]: I0321 08:29:27.534278 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:27 crc kubenswrapper[4696]: E0321 08:29:27.534553 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 08:29:29 crc kubenswrapper[4696]: I0321 08:29:29.533833 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:29 crc kubenswrapper[4696]: I0321 08:29:29.533868 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:29 crc kubenswrapper[4696]: E0321 08:29:29.533943 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 08:29:29 crc kubenswrapper[4696]: I0321 08:29:29.533892 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:29 crc kubenswrapper[4696]: E0321 08:29:29.534109 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 08:29:29 crc kubenswrapper[4696]: E0321 08:29:29.534189 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 08:29:29 crc kubenswrapper[4696]: I0321 08:29:29.991256 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-phrtr"] Mar 21 08:29:29 crc kubenswrapper[4696]: I0321 08:29:29.991573 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-phrtr" Mar 21 08:29:29 crc kubenswrapper[4696]: I0321 08:29:29.993680 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 21 08:29:29 crc kubenswrapper[4696]: I0321 08:29:29.994269 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 21 08:29:29 crc kubenswrapper[4696]: I0321 08:29:29.995023 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.009208 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-t6d75"] Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.009753 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-4rp52"] Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.010339 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-z7srw"] Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.010709 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.011098 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.011546 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.014047 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.014109 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.014360 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.014482 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.014613 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.014886 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.014634 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.014977 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.014677 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.014697 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.014767 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.016119 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.019520 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9tp76"] Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.020451 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.022672 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.022912 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.023266 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.026354 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.026809 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.027024 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.027469 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.048505 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-var-lib-openvswitch\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.048547 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-system-cni-dir\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.048572 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1472d11d-4da5-4423-99b1-4dd73e3e9b9b-cnibin\") pod \"multus-additional-cni-plugins-4rp52\" (UID: \"1472d11d-4da5-4423-99b1-4dd73e3e9b9b\") " pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.048597 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-run-netns\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.048613 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-host-run-netns\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.048629 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/daaf227a-2305-495e-8495-a280abcd8e10-mcd-auth-proxy-config\") pod \"machine-config-daemon-z7srw\" (UID: \"daaf227a-2305-495e-8495-a280abcd8e10\") " pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.048647 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1472d11d-4da5-4423-99b1-4dd73e3e9b9b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-4rp52\" (UID: \"1472d11d-4da5-4423-99b1-4dd73e3e9b9b\") " pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.048664 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-kubelet\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.048683 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-cni-bin\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.048701 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1472d11d-4da5-4423-99b1-4dd73e3e9b9b-system-cni-dir\") pod \"multus-additional-cni-plugins-4rp52\" (UID: \"1472d11d-4da5-4423-99b1-4dd73e3e9b9b\") " pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.048720 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/42c712ed-b850-4947-b16e-5517c26612f8-multus-daemon-config\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.048740 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-etc-kubernetes\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.048755 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-etc-openvswitch\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.048771 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-cni-netd\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.048786 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9f0e7707-cb43-4f15-b309-7f651da35ddf-ovn-node-metrics-cert\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.048801 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-os-release\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.048888 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-multus-socket-dir-parent\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.048919 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-host-run-k8s-cni-cncf-io\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.048935 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-host-var-lib-cni-multus\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.048951 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/daaf227a-2305-495e-8495-a280abcd8e10-proxy-tls\") pod \"machine-config-daemon-z7srw\" (UID: \"daaf227a-2305-495e-8495-a280abcd8e10\") " pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049008 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-hostroot\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049035 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-host-run-multus-certs\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049055 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mxmf\" (UniqueName: \"kubernetes.io/projected/42c712ed-b850-4947-b16e-5517c26612f8-kube-api-access-4mxmf\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049073 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-node-log\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049086 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-run-openvswitch\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049100 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-multus-cni-dir\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049113 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/42c712ed-b850-4947-b16e-5517c26612f8-cni-binary-copy\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049279 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvvxm\" (UniqueName: \"kubernetes.io/projected/1472d11d-4da5-4423-99b1-4dd73e3e9b9b-kube-api-access-tvvxm\") pod \"multus-additional-cni-plugins-4rp52\" (UID: \"1472d11d-4da5-4423-99b1-4dd73e3e9b9b\") " pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049296 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-run-ovn-kubernetes\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049317 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9f0e7707-cb43-4f15-b309-7f651da35ddf-env-overrides\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049337 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1472d11d-4da5-4423-99b1-4dd73e3e9b9b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-4rp52\" (UID: \"1472d11d-4da5-4423-99b1-4dd73e3e9b9b\") " pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049356 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-run-ovn\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049371 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-host-var-lib-kubelet\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049399 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/16d55437-4d46-4a91-bc54-7654442439bc-hosts-file\") pod \"node-resolver-phrtr\" (UID: \"16d55437-4d46-4a91-bc54-7654442439bc\") " pod="openshift-dns/node-resolver-phrtr" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049420 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1472d11d-4da5-4423-99b1-4dd73e3e9b9b-cni-binary-copy\") pod \"multus-additional-cni-plugins-4rp52\" (UID: \"1472d11d-4da5-4423-99b1-4dd73e3e9b9b\") " pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049456 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-run-systemd\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049488 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049511 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9f0e7707-cb43-4f15-b309-7f651da35ddf-ovnkube-script-lib\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049554 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1472d11d-4da5-4423-99b1-4dd73e3e9b9b-os-release\") pod \"multus-additional-cni-plugins-4rp52\" (UID: \"1472d11d-4da5-4423-99b1-4dd73e3e9b9b\") " pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049578 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-systemd-units\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049597 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-slash\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049629 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bw42\" (UniqueName: \"kubernetes.io/projected/16d55437-4d46-4a91-bc54-7654442439bc-kube-api-access-7bw42\") pod \"node-resolver-phrtr\" (UID: \"16d55437-4d46-4a91-bc54-7654442439bc\") " pod="openshift-dns/node-resolver-phrtr" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049652 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9f0e7707-cb43-4f15-b309-7f651da35ddf-ovnkube-config\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049670 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stzpr\" (UniqueName: \"kubernetes.io/projected/9f0e7707-cb43-4f15-b309-7f651da35ddf-kube-api-access-stzpr\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049691 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-multus-conf-dir\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049712 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-log-socket\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049732 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-cnibin\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049751 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/daaf227a-2305-495e-8495-a280abcd8e10-rootfs\") pod \"machine-config-daemon-z7srw\" (UID: \"daaf227a-2305-495e-8495-a280abcd8e10\") " pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049773 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4z7l\" (UniqueName: \"kubernetes.io/projected/daaf227a-2305-495e-8495-a280abcd8e10-kube-api-access-w4z7l\") pod \"machine-config-daemon-z7srw\" (UID: \"daaf227a-2305-495e-8495-a280abcd8e10\") " pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.049796 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-host-var-lib-cni-bin\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.079845 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-zjrl8"] Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.080273 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zjrl8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.082634 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.082694 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.082729 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.082810 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151265 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-systemd-units\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151307 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-slash\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151328 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bw42\" (UniqueName: \"kubernetes.io/projected/16d55437-4d46-4a91-bc54-7654442439bc-kube-api-access-7bw42\") pod \"node-resolver-phrtr\" (UID: \"16d55437-4d46-4a91-bc54-7654442439bc\") " pod="openshift-dns/node-resolver-phrtr" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151344 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1472d11d-4da5-4423-99b1-4dd73e3e9b9b-os-release\") pod \"multus-additional-cni-plugins-4rp52\" (UID: \"1472d11d-4da5-4423-99b1-4dd73e3e9b9b\") " pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151359 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9f0e7707-cb43-4f15-b309-7f651da35ddf-ovnkube-config\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151380 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stzpr\" (UniqueName: \"kubernetes.io/projected/9f0e7707-cb43-4f15-b309-7f651da35ddf-kube-api-access-stzpr\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151399 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-multus-conf-dir\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151419 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-log-socket\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151425 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-systemd-units\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151509 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1472d11d-4da5-4423-99b1-4dd73e3e9b9b-os-release\") pod \"multus-additional-cni-plugins-4rp52\" (UID: \"1472d11d-4da5-4423-99b1-4dd73e3e9b9b\") " pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151513 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-cnibin\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151441 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-cnibin\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151563 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4z7l\" (UniqueName: \"kubernetes.io/projected/daaf227a-2305-495e-8495-a280abcd8e10-kube-api-access-w4z7l\") pod \"machine-config-daemon-z7srw\" (UID: \"daaf227a-2305-495e-8495-a280abcd8e10\") " pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151596 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-host-var-lib-cni-bin\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151617 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-slash\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151645 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/daaf227a-2305-495e-8495-a280abcd8e10-rootfs\") pod \"machine-config-daemon-z7srw\" (UID: \"daaf227a-2305-495e-8495-a280abcd8e10\") " pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151620 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/daaf227a-2305-495e-8495-a280abcd8e10-rootfs\") pod \"machine-config-daemon-z7srw\" (UID: \"daaf227a-2305-495e-8495-a280abcd8e10\") " pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151672 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-var-lib-openvswitch\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151686 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-system-cni-dir\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151703 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/849f5514-02ae-4a53-a5e0-ad65689b7d6d-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-zjrl8\" (UID: \"849f5514-02ae-4a53-a5e0-ad65689b7d6d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zjrl8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151720 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1472d11d-4da5-4423-99b1-4dd73e3e9b9b-cnibin\") pod \"multus-additional-cni-plugins-4rp52\" (UID: \"1472d11d-4da5-4423-99b1-4dd73e3e9b9b\") " pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151733 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-run-netns\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151747 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-host-run-netns\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151760 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/daaf227a-2305-495e-8495-a280abcd8e10-mcd-auth-proxy-config\") pod \"machine-config-daemon-z7srw\" (UID: \"daaf227a-2305-495e-8495-a280abcd8e10\") " pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151775 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-kubelet\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151788 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-cni-bin\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151804 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/849f5514-02ae-4a53-a5e0-ad65689b7d6d-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-zjrl8\" (UID: \"849f5514-02ae-4a53-a5e0-ad65689b7d6d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zjrl8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151844 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1472d11d-4da5-4423-99b1-4dd73e3e9b9b-system-cni-dir\") pod \"multus-additional-cni-plugins-4rp52\" (UID: \"1472d11d-4da5-4423-99b1-4dd73e3e9b9b\") " pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151863 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1472d11d-4da5-4423-99b1-4dd73e3e9b9b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-4rp52\" (UID: \"1472d11d-4da5-4423-99b1-4dd73e3e9b9b\") " pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151882 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-etc-kubernetes\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151900 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-etc-openvswitch\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151906 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-multus-conf-dir\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151915 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-cni-netd\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151931 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9f0e7707-cb43-4f15-b309-7f651da35ddf-ovn-node-metrics-cert\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151933 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-log-socket\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151946 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/42c712ed-b850-4947-b16e-5517c26612f8-multus-daemon-config\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151958 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1472d11d-4da5-4423-99b1-4dd73e3e9b9b-cnibin\") pod \"multus-additional-cni-plugins-4rp52\" (UID: \"1472d11d-4da5-4423-99b1-4dd73e3e9b9b\") " pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151957 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-kubelet\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151964 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-multus-socket-dir-parent\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151995 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-multus-socket-dir-parent\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151995 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-var-lib-openvswitch\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152006 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-host-run-k8s-cni-cncf-io\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152017 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-cni-bin\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152031 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-system-cni-dir\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152033 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-host-var-lib-cni-multus\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152043 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1472d11d-4da5-4423-99b1-4dd73e3e9b9b-system-cni-dir\") pod \"multus-additional-cni-plugins-4rp52\" (UID: \"1472d11d-4da5-4423-99b1-4dd73e3e9b9b\") " pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152056 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/daaf227a-2305-495e-8495-a280abcd8e10-proxy-tls\") pod \"machine-config-daemon-z7srw\" (UID: \"daaf227a-2305-495e-8495-a280abcd8e10\") " pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152065 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-host-run-netns\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152089 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-os-release\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152107 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-etc-openvswitch\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152113 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-host-run-multus-certs\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152128 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-etc-kubernetes\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.151882 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-host-var-lib-cni-bin\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152136 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mxmf\" (UniqueName: \"kubernetes.io/projected/42c712ed-b850-4947-b16e-5517c26612f8-kube-api-access-4mxmf\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152163 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-node-log\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152178 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-hostroot\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152193 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-multus-cni-dir\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152208 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/42c712ed-b850-4947-b16e-5517c26612f8-cni-binary-copy\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152230 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvvxm\" (UniqueName: \"kubernetes.io/projected/1472d11d-4da5-4423-99b1-4dd73e3e9b9b-kube-api-access-tvvxm\") pod \"multus-additional-cni-plugins-4rp52\" (UID: \"1472d11d-4da5-4423-99b1-4dd73e3e9b9b\") " pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152244 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-run-openvswitch\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152260 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-run-ovn-kubernetes\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152280 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/849f5514-02ae-4a53-a5e0-ad65689b7d6d-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-zjrl8\" (UID: \"849f5514-02ae-4a53-a5e0-ad65689b7d6d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zjrl8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152299 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9f0e7707-cb43-4f15-b309-7f651da35ddf-env-overrides\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152316 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/849f5514-02ae-4a53-a5e0-ad65689b7d6d-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-zjrl8\" (UID: \"849f5514-02ae-4a53-a5e0-ad65689b7d6d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zjrl8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152333 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-run-ovn\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152334 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9f0e7707-cb43-4f15-b309-7f651da35ddf-ovnkube-config\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152349 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-host-var-lib-kubelet\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152385 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/16d55437-4d46-4a91-bc54-7654442439bc-hosts-file\") pod \"node-resolver-phrtr\" (UID: \"16d55437-4d46-4a91-bc54-7654442439bc\") " pod="openshift-dns/node-resolver-phrtr" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152404 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1472d11d-4da5-4423-99b1-4dd73e3e9b9b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-4rp52\" (UID: \"1472d11d-4da5-4423-99b1-4dd73e3e9b9b\") " pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152422 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1472d11d-4da5-4423-99b1-4dd73e3e9b9b-cni-binary-copy\") pod \"multus-additional-cni-plugins-4rp52\" (UID: \"1472d11d-4da5-4423-99b1-4dd73e3e9b9b\") " pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152444 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-run-systemd\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152467 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152488 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9f0e7707-cb43-4f15-b309-7f651da35ddf-ovnkube-script-lib\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152523 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/849f5514-02ae-4a53-a5e0-ad65689b7d6d-service-ca\") pod \"cluster-version-operator-5c965bbfc6-zjrl8\" (UID: \"849f5514-02ae-4a53-a5e0-ad65689b7d6d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zjrl8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152577 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-node-log\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152600 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-hostroot\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152639 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-multus-cni-dir\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152661 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1472d11d-4da5-4423-99b1-4dd73e3e9b9b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-4rp52\" (UID: \"1472d11d-4da5-4423-99b1-4dd73e3e9b9b\") " pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152694 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-cni-netd\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152930 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/daaf227a-2305-495e-8495-a280abcd8e10-mcd-auth-proxy-config\") pod \"machine-config-daemon-z7srw\" (UID: \"daaf227a-2305-495e-8495-a280abcd8e10\") " pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152996 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-os-release\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.152091 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-run-netns\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.153045 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-host-run-multus-certs\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.153078 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-run-ovn\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.153251 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-run-openvswitch\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.153286 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-run-ovn-kubernetes\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.153288 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/42c712ed-b850-4947-b16e-5517c26612f8-cni-binary-copy\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.153310 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-host-var-lib-kubelet\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.153351 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/16d55437-4d46-4a91-bc54-7654442439bc-hosts-file\") pod \"node-resolver-phrtr\" (UID: \"16d55437-4d46-4a91-bc54-7654442439bc\") " pod="openshift-dns/node-resolver-phrtr" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.153724 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9f0e7707-cb43-4f15-b309-7f651da35ddf-env-overrides\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.153867 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1472d11d-4da5-4423-99b1-4dd73e3e9b9b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-4rp52\" (UID: \"1472d11d-4da5-4423-99b1-4dd73e3e9b9b\") " pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.153957 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.154013 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-host-run-k8s-cni-cncf-io\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.154030 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-run-systemd\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.154056 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/42c712ed-b850-4947-b16e-5517c26612f8-host-var-lib-cni-multus\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.154305 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/42c712ed-b850-4947-b16e-5517c26612f8-multus-daemon-config\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.154319 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1472d11d-4da5-4423-99b1-4dd73e3e9b9b-cni-binary-copy\") pod \"multus-additional-cni-plugins-4rp52\" (UID: \"1472d11d-4da5-4423-99b1-4dd73e3e9b9b\") " pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.154797 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9f0e7707-cb43-4f15-b309-7f651da35ddf-ovnkube-script-lib\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.157547 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/daaf227a-2305-495e-8495-a280abcd8e10-proxy-tls\") pod \"machine-config-daemon-z7srw\" (UID: \"daaf227a-2305-495e-8495-a280abcd8e10\") " pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.159651 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9f0e7707-cb43-4f15-b309-7f651da35ddf-ovn-node-metrics-cert\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.166363 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-wzrdm"] Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.166841 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-wzrdm" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.169068 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.169216 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.169339 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.172139 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.172309 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4z7l\" (UniqueName: \"kubernetes.io/projected/daaf227a-2305-495e-8495-a280abcd8e10-kube-api-access-w4z7l\") pod \"machine-config-daemon-z7srw\" (UID: \"daaf227a-2305-495e-8495-a280abcd8e10\") " pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.173901 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bw42\" (UniqueName: \"kubernetes.io/projected/16d55437-4d46-4a91-bc54-7654442439bc-kube-api-access-7bw42\") pod \"node-resolver-phrtr\" (UID: \"16d55437-4d46-4a91-bc54-7654442439bc\") " pod="openshift-dns/node-resolver-phrtr" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.181834 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvvxm\" (UniqueName: \"kubernetes.io/projected/1472d11d-4da5-4423-99b1-4dd73e3e9b9b-kube-api-access-tvvxm\") pod \"multus-additional-cni-plugins-4rp52\" (UID: \"1472d11d-4da5-4423-99b1-4dd73e3e9b9b\") " pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.181849 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mxmf\" (UniqueName: \"kubernetes.io/projected/42c712ed-b850-4947-b16e-5517c26612f8-kube-api-access-4mxmf\") pod \"multus-t6d75\" (UID: \"42c712ed-b850-4947-b16e-5517c26612f8\") " pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.184040 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stzpr\" (UniqueName: \"kubernetes.io/projected/9f0e7707-cb43-4f15-b309-7f651da35ddf-kube-api-access-stzpr\") pod \"ovnkube-node-9tp76\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.253136 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/849f5514-02ae-4a53-a5e0-ad65689b7d6d-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-zjrl8\" (UID: \"849f5514-02ae-4a53-a5e0-ad65689b7d6d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zjrl8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.253484 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/849f5514-02ae-4a53-a5e0-ad65689b7d6d-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-zjrl8\" (UID: \"849f5514-02ae-4a53-a5e0-ad65689b7d6d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zjrl8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.253576 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/849f5514-02ae-4a53-a5e0-ad65689b7d6d-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-zjrl8\" (UID: \"849f5514-02ae-4a53-a5e0-ad65689b7d6d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zjrl8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.253720 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/849f5514-02ae-4a53-a5e0-ad65689b7d6d-service-ca\") pod \"cluster-version-operator-5c965bbfc6-zjrl8\" (UID: \"849f5514-02ae-4a53-a5e0-ad65689b7d6d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zjrl8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.253858 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/15bb2a3c-d497-4673-aae2-fe93e84d4bed-host\") pod \"node-ca-wzrdm\" (UID: \"15bb2a3c-d497-4673-aae2-fe93e84d4bed\") " pod="openshift-image-registry/node-ca-wzrdm" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.253977 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/849f5514-02ae-4a53-a5e0-ad65689b7d6d-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-zjrl8\" (UID: \"849f5514-02ae-4a53-a5e0-ad65689b7d6d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zjrl8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.254098 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/849f5514-02ae-4a53-a5e0-ad65689b7d6d-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-zjrl8\" (UID: \"849f5514-02ae-4a53-a5e0-ad65689b7d6d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zjrl8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.254218 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkfsg\" (UniqueName: \"kubernetes.io/projected/15bb2a3c-d497-4673-aae2-fe93e84d4bed-kube-api-access-qkfsg\") pod \"node-ca-wzrdm\" (UID: \"15bb2a3c-d497-4673-aae2-fe93e84d4bed\") " pod="openshift-image-registry/node-ca-wzrdm" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.254318 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/15bb2a3c-d497-4673-aae2-fe93e84d4bed-serviceca\") pod \"node-ca-wzrdm\" (UID: \"15bb2a3c-d497-4673-aae2-fe93e84d4bed\") " pod="openshift-image-registry/node-ca-wzrdm" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.254418 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/849f5514-02ae-4a53-a5e0-ad65689b7d6d-service-ca\") pod \"cluster-version-operator-5c965bbfc6-zjrl8\" (UID: \"849f5514-02ae-4a53-a5e0-ad65689b7d6d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zjrl8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.254234 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/849f5514-02ae-4a53-a5e0-ad65689b7d6d-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-zjrl8\" (UID: \"849f5514-02ae-4a53-a5e0-ad65689b7d6d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zjrl8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.256848 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/849f5514-02ae-4a53-a5e0-ad65689b7d6d-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-zjrl8\" (UID: \"849f5514-02ae-4a53-a5e0-ad65689b7d6d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zjrl8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.268148 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/849f5514-02ae-4a53-a5e0-ad65689b7d6d-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-zjrl8\" (UID: \"849f5514-02ae-4a53-a5e0-ad65689b7d6d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zjrl8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.292419 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5n6t8"] Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.292912 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5n6t8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.294191 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.298155 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.309796 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-phrtr" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.316056 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-w95sn"] Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.316424 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w95sn" Mar 21 08:29:30 crc kubenswrapper[4696]: E0321 08:29:30.316481 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w95sn" podUID="9bbfb96a-607b-47df-83df-7c75dd94ad71" Mar 21 08:29:30 crc kubenswrapper[4696]: W0321 08:29:30.321845 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16d55437_4d46_4a91_bc54_7654442439bc.slice/crio-c0c5d98404e3af732d6e8fe8e537f26a7b978f8caa59c26f3481b5485a7e4a19 WatchSource:0}: Error finding container c0c5d98404e3af732d6e8fe8e537f26a7b978f8caa59c26f3481b5485a7e4a19: Status 404 returned error can't find the container with id c0c5d98404e3af732d6e8fe8e537f26a7b978f8caa59c26f3481b5485a7e4a19 Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.340306 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.347577 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-t6d75" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.355059 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/15bb2a3c-d497-4673-aae2-fe93e84d4bed-serviceca\") pod \"node-ca-wzrdm\" (UID: \"15bb2a3c-d497-4673-aae2-fe93e84d4bed\") " pod="openshift-image-registry/node-ca-wzrdm" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.355124 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d4636ac5-84e3-4a74-9ea0-c399ff7554d1-env-overrides\") pod \"ovnkube-control-plane-749d76644c-5n6t8\" (UID: \"d4636ac5-84e3-4a74-9ea0-c399ff7554d1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5n6t8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.355146 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/15bb2a3c-d497-4673-aae2-fe93e84d4bed-host\") pod \"node-ca-wzrdm\" (UID: \"15bb2a3c-d497-4673-aae2-fe93e84d4bed\") " pod="openshift-image-registry/node-ca-wzrdm" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.355164 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d4636ac5-84e3-4a74-9ea0-c399ff7554d1-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-5n6t8\" (UID: \"d4636ac5-84e3-4a74-9ea0-c399ff7554d1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5n6t8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.355181 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5dwd\" (UniqueName: \"kubernetes.io/projected/9bbfb96a-607b-47df-83df-7c75dd94ad71-kube-api-access-x5dwd\") pod \"network-metrics-daemon-w95sn\" (UID: \"9bbfb96a-607b-47df-83df-7c75dd94ad71\") " pod="openshift-multus/network-metrics-daemon-w95sn" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.355197 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d4636ac5-84e3-4a74-9ea0-c399ff7554d1-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-5n6t8\" (UID: \"d4636ac5-84e3-4a74-9ea0-c399ff7554d1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5n6t8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.355215 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9bbfb96a-607b-47df-83df-7c75dd94ad71-metrics-certs\") pod \"network-metrics-daemon-w95sn\" (UID: \"9bbfb96a-607b-47df-83df-7c75dd94ad71\") " pod="openshift-multus/network-metrics-daemon-w95sn" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.355234 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d59lg\" (UniqueName: \"kubernetes.io/projected/d4636ac5-84e3-4a74-9ea0-c399ff7554d1-kube-api-access-d59lg\") pod \"ovnkube-control-plane-749d76644c-5n6t8\" (UID: \"d4636ac5-84e3-4a74-9ea0-c399ff7554d1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5n6t8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.355261 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkfsg\" (UniqueName: \"kubernetes.io/projected/15bb2a3c-d497-4673-aae2-fe93e84d4bed-kube-api-access-qkfsg\") pod \"node-ca-wzrdm\" (UID: \"15bb2a3c-d497-4673-aae2-fe93e84d4bed\") " pod="openshift-image-registry/node-ca-wzrdm" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.356483 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-4rp52" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.356618 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/15bb2a3c-d497-4673-aae2-fe93e84d4bed-host\") pod \"node-ca-wzrdm\" (UID: \"15bb2a3c-d497-4673-aae2-fe93e84d4bed\") " pod="openshift-image-registry/node-ca-wzrdm" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.358160 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/15bb2a3c-d497-4673-aae2-fe93e84d4bed-serviceca\") pod \"node-ca-wzrdm\" (UID: \"15bb2a3c-d497-4673-aae2-fe93e84d4bed\") " pod="openshift-image-registry/node-ca-wzrdm" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.361279 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:30 crc kubenswrapper[4696]: W0321 08:29:30.361646 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42c712ed_b850_4947_b16e_5517c26612f8.slice/crio-6c79fab98b1f2aa475d8679d897ebbc9978586b3fca9c9deac9734ed1dbda603 WatchSource:0}: Error finding container 6c79fab98b1f2aa475d8679d897ebbc9978586b3fca9c9deac9734ed1dbda603: Status 404 returned error can't find the container with id 6c79fab98b1f2aa475d8679d897ebbc9978586b3fca9c9deac9734ed1dbda603 Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.371193 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkfsg\" (UniqueName: \"kubernetes.io/projected/15bb2a3c-d497-4673-aae2-fe93e84d4bed-kube-api-access-qkfsg\") pod \"node-ca-wzrdm\" (UID: \"15bb2a3c-d497-4673-aae2-fe93e84d4bed\") " pod="openshift-image-registry/node-ca-wzrdm" Mar 21 08:29:30 crc kubenswrapper[4696]: W0321 08:29:30.377456 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1472d11d_4da5_4423_99b1_4dd73e3e9b9b.slice/crio-95d01d88a06a1419c4d56d2e4023de25df8a5db2cd993f3b3465f3cd9e6afb19 WatchSource:0}: Error finding container 95d01d88a06a1419c4d56d2e4023de25df8a5db2cd993f3b3465f3cd9e6afb19: Status 404 returned error can't find the container with id 95d01d88a06a1419c4d56d2e4023de25df8a5db2cd993f3b3465f3cd9e6afb19 Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.392925 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zjrl8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.456456 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d4636ac5-84e3-4a74-9ea0-c399ff7554d1-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-5n6t8\" (UID: \"d4636ac5-84e3-4a74-9ea0-c399ff7554d1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5n6t8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.456499 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5dwd\" (UniqueName: \"kubernetes.io/projected/9bbfb96a-607b-47df-83df-7c75dd94ad71-kube-api-access-x5dwd\") pod \"network-metrics-daemon-w95sn\" (UID: \"9bbfb96a-607b-47df-83df-7c75dd94ad71\") " pod="openshift-multus/network-metrics-daemon-w95sn" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.456569 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d4636ac5-84e3-4a74-9ea0-c399ff7554d1-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-5n6t8\" (UID: \"d4636ac5-84e3-4a74-9ea0-c399ff7554d1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5n6t8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.456589 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9bbfb96a-607b-47df-83df-7c75dd94ad71-metrics-certs\") pod \"network-metrics-daemon-w95sn\" (UID: \"9bbfb96a-607b-47df-83df-7c75dd94ad71\") " pod="openshift-multus/network-metrics-daemon-w95sn" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.456607 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d59lg\" (UniqueName: \"kubernetes.io/projected/d4636ac5-84e3-4a74-9ea0-c399ff7554d1-kube-api-access-d59lg\") pod \"ovnkube-control-plane-749d76644c-5n6t8\" (UID: \"d4636ac5-84e3-4a74-9ea0-c399ff7554d1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5n6t8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.456668 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d4636ac5-84e3-4a74-9ea0-c399ff7554d1-env-overrides\") pod \"ovnkube-control-plane-749d76644c-5n6t8\" (UID: \"d4636ac5-84e3-4a74-9ea0-c399ff7554d1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5n6t8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.457216 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d4636ac5-84e3-4a74-9ea0-c399ff7554d1-env-overrides\") pod \"ovnkube-control-plane-749d76644c-5n6t8\" (UID: \"d4636ac5-84e3-4a74-9ea0-c399ff7554d1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5n6t8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.457597 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d4636ac5-84e3-4a74-9ea0-c399ff7554d1-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-5n6t8\" (UID: \"d4636ac5-84e3-4a74-9ea0-c399ff7554d1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5n6t8" Mar 21 08:29:30 crc kubenswrapper[4696]: E0321 08:29:30.458451 4696 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 08:29:30 crc kubenswrapper[4696]: E0321 08:29:30.458500 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9bbfb96a-607b-47df-83df-7c75dd94ad71-metrics-certs podName:9bbfb96a-607b-47df-83df-7c75dd94ad71 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:30.958488713 +0000 UTC m=+105.079369426 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9bbfb96a-607b-47df-83df-7c75dd94ad71-metrics-certs") pod "network-metrics-daemon-w95sn" (UID: "9bbfb96a-607b-47df-83df-7c75dd94ad71") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.462164 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d4636ac5-84e3-4a74-9ea0-c399ff7554d1-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-5n6t8\" (UID: \"d4636ac5-84e3-4a74-9ea0-c399ff7554d1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5n6t8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.473285 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d59lg\" (UniqueName: \"kubernetes.io/projected/d4636ac5-84e3-4a74-9ea0-c399ff7554d1-kube-api-access-d59lg\") pod \"ovnkube-control-plane-749d76644c-5n6t8\" (UID: \"d4636ac5-84e3-4a74-9ea0-c399ff7554d1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5n6t8" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.474664 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5dwd\" (UniqueName: \"kubernetes.io/projected/9bbfb96a-607b-47df-83df-7c75dd94ad71-kube-api-access-x5dwd\") pod \"network-metrics-daemon-w95sn\" (UID: \"9bbfb96a-607b-47df-83df-7c75dd94ad71\") " pod="openshift-multus/network-metrics-daemon-w95sn" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.495320 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-wzrdm" Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.535040 4696 scope.go:117] "RemoveContainer" containerID="a78fedb8a70e8b5cec10a972e63964e900746253f6add19cbedf2e47af762244" Mar 21 08:29:30 crc kubenswrapper[4696]: W0321 08:29:30.549919 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15bb2a3c_d497_4673_aae2_fe93e84d4bed.slice/crio-1d233d5ea3dd16fd025d66464e9ec9133d4fbeba34a66903d67352ddcc052d14 WatchSource:0}: Error finding container 1d233d5ea3dd16fd025d66464e9ec9133d4fbeba34a66903d67352ddcc052d14: Status 404 returned error can't find the container with id 1d233d5ea3dd16fd025d66464e9ec9133d4fbeba34a66903d67352ddcc052d14 Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.604248 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5n6t8" Mar 21 08:29:30 crc kubenswrapper[4696]: W0321 08:29:30.623930 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4636ac5_84e3_4a74_9ea0_c399ff7554d1.slice/crio-73d4804f36942e5ab3d6b8eeacef9fb4928c1f573c18a950f168ece8a786ddda WatchSource:0}: Error finding container 73d4804f36942e5ab3d6b8eeacef9fb4928c1f573c18a950f168ece8a786ddda: Status 404 returned error can't find the container with id 73d4804f36942e5ab3d6b8eeacef9fb4928c1f573c18a950f168ece8a786ddda Mar 21 08:29:30 crc kubenswrapper[4696]: I0321 08:29:30.961650 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9bbfb96a-607b-47df-83df-7c75dd94ad71-metrics-certs\") pod \"network-metrics-daemon-w95sn\" (UID: \"9bbfb96a-607b-47df-83df-7c75dd94ad71\") " pod="openshift-multus/network-metrics-daemon-w95sn" Mar 21 08:29:30 crc kubenswrapper[4696]: E0321 08:29:30.961807 4696 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 08:29:30 crc kubenswrapper[4696]: E0321 08:29:30.961886 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9bbfb96a-607b-47df-83df-7c75dd94ad71-metrics-certs podName:9bbfb96a-607b-47df-83df-7c75dd94ad71 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:31.961870737 +0000 UTC m=+106.082751450 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9bbfb96a-607b-47df-83df-7c75dd94ad71-metrics-certs") pod "network-metrics-daemon-w95sn" (UID: "9bbfb96a-607b-47df-83df-7c75dd94ad71") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.006325 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5n6t8" event={"ID":"d4636ac5-84e3-4a74-9ea0-c399ff7554d1","Type":"ContainerStarted","Data":"6a421524ba6285bd927311ca82107e8506a7bac35b71e0b0bc4f014d8087de80"} Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.006381 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5n6t8" event={"ID":"d4636ac5-84e3-4a74-9ea0-c399ff7554d1","Type":"ContainerStarted","Data":"64d5278134111f2bd516a682ce250ebb3f7c002dfb8a6b60564a082fdc9fc5f2"} Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.006392 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5n6t8" event={"ID":"d4636ac5-84e3-4a74-9ea0-c399ff7554d1","Type":"ContainerStarted","Data":"73d4804f36942e5ab3d6b8eeacef9fb4928c1f573c18a950f168ece8a786ddda"} Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.010368 4696 generic.go:334] "Generic (PLEG): container finished" podID="1472d11d-4da5-4423-99b1-4dd73e3e9b9b" containerID="a06f9629511006cecbbaea849b0120d82cd73f2924a97142e63caf7d60186d0a" exitCode=0 Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.010424 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4rp52" event={"ID":"1472d11d-4da5-4423-99b1-4dd73e3e9b9b","Type":"ContainerDied","Data":"a06f9629511006cecbbaea849b0120d82cd73f2924a97142e63caf7d60186d0a"} Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.010487 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4rp52" event={"ID":"1472d11d-4da5-4423-99b1-4dd73e3e9b9b","Type":"ContainerStarted","Data":"95d01d88a06a1419c4d56d2e4023de25df8a5db2cd993f3b3465f3cd9e6afb19"} Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.012901 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.014542 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"815425653c2afb43bac59c2ba2c04bfe71edde9aa3372e516741e02a9745bccd"} Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.015033 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.016522 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-wzrdm" event={"ID":"15bb2a3c-d497-4673-aae2-fe93e84d4bed","Type":"ContainerStarted","Data":"22157c3a08bfcfebd660f2f1246670f03f72fc475e43786a8ca1af64cb49d2b2"} Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.016549 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-wzrdm" event={"ID":"15bb2a3c-d497-4673-aae2-fe93e84d4bed","Type":"ContainerStarted","Data":"1d233d5ea3dd16fd025d66464e9ec9133d4fbeba34a66903d67352ddcc052d14"} Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.018309 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-phrtr" event={"ID":"16d55437-4d46-4a91-bc54-7654442439bc","Type":"ContainerStarted","Data":"cfba95b2f38a230afed3ff95a2de2627f52581b2e24223c07c8a97beab16d698"} Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.018342 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-phrtr" event={"ID":"16d55437-4d46-4a91-bc54-7654442439bc","Type":"ContainerStarted","Data":"c0c5d98404e3af732d6e8fe8e537f26a7b978f8caa59c26f3481b5485a7e4a19"} Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.023584 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5n6t8" podStartSLOduration=64.023568362 podStartE2EDuration="1m4.023568362s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:31.022502104 +0000 UTC m=+105.143382827" watchObservedRunningTime="2026-03-21 08:29:31.023568362 +0000 UTC m=+105.144449075" Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.026747 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-t6d75" event={"ID":"42c712ed-b850-4947-b16e-5517c26612f8","Type":"ContainerStarted","Data":"c4d2a77be0485142291c9c1ec006aac3d018189240bc4f94c8f175b4cf291b8f"} Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.026796 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-t6d75" event={"ID":"42c712ed-b850-4947-b16e-5517c26612f8","Type":"ContainerStarted","Data":"6c79fab98b1f2aa475d8679d897ebbc9978586b3fca9c9deac9734ed1dbda603"} Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.028560 4696 generic.go:334] "Generic (PLEG): container finished" podID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerID="af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c" exitCode=0 Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.028619 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" event={"ID":"9f0e7707-cb43-4f15-b309-7f651da35ddf","Type":"ContainerDied","Data":"af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c"} Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.028645 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" event={"ID":"9f0e7707-cb43-4f15-b309-7f651da35ddf","Type":"ContainerStarted","Data":"a539a3280d13394f870f8a68e82bffd19b597121e60aeb43f584fe392860c830"} Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.030809 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerStarted","Data":"cc03f672f40918eb6b20d28fca5ea86f66b82b5edb78501baa6f6498ea5b706f"} Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.030877 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerStarted","Data":"9606dfcaf248d5014358578adf21a000d5fb5d698e3980a6ae0ec673dfb6ef85"} Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.030891 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerStarted","Data":"eef8c3888247b5355e768431411b8df3c6484ca01d7d3b20b4f5a2ab44fb8bf3"} Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.035333 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zjrl8" event={"ID":"849f5514-02ae-4a53-a5e0-ad65689b7d6d","Type":"ContainerStarted","Data":"d4dd61636ad57e307f5cf41bbaffe1577d32c4552bac5016f5a3aeb7e74e226c"} Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.035386 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zjrl8" event={"ID":"849f5514-02ae-4a53-a5e0-ad65689b7d6d","Type":"ContainerStarted","Data":"bfa3942659a86fd8eb09eb6634768dbbc24594a2fd71e35bc1f42fc896adb510"} Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.037884 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-wzrdm" podStartSLOduration=65.037872914 podStartE2EDuration="1m5.037872914s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:31.037591747 +0000 UTC m=+105.158472470" watchObservedRunningTime="2026-03-21 08:29:31.037872914 +0000 UTC m=+105.158753627" Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.094688 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=13.094674032 podStartE2EDuration="13.094674032s" podCreationTimestamp="2026-03-21 08:29:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:31.094289082 +0000 UTC m=+105.215169815" watchObservedRunningTime="2026-03-21 08:29:31.094674032 +0000 UTC m=+105.215554745" Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.135904 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-phrtr" podStartSLOduration=65.135882594 podStartE2EDuration="1m5.135882594s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:31.108696757 +0000 UTC m=+105.229577470" watchObservedRunningTime="2026-03-21 08:29:31.135882594 +0000 UTC m=+105.256763307" Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.148540 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zjrl8" podStartSLOduration=65.148523633 podStartE2EDuration="1m5.148523633s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:31.148108462 +0000 UTC m=+105.268989195" watchObservedRunningTime="2026-03-21 08:29:31.148523633 +0000 UTC m=+105.269404356" Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.168253 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podStartSLOduration=65.168239215 podStartE2EDuration="1m5.168239215s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:31.167946278 +0000 UTC m=+105.288826981" watchObservedRunningTime="2026-03-21 08:29:31.168239215 +0000 UTC m=+105.289119928" Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.533906 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.533959 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.533912 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:31 crc kubenswrapper[4696]: E0321 08:29:31.534023 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 08:29:31 crc kubenswrapper[4696]: E0321 08:29:31.534088 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 08:29:31 crc kubenswrapper[4696]: E0321 08:29:31.534208 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 08:29:31 crc kubenswrapper[4696]: I0321 08:29:31.974290 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9bbfb96a-607b-47df-83df-7c75dd94ad71-metrics-certs\") pod \"network-metrics-daemon-w95sn\" (UID: \"9bbfb96a-607b-47df-83df-7c75dd94ad71\") " pod="openshift-multus/network-metrics-daemon-w95sn" Mar 21 08:29:31 crc kubenswrapper[4696]: E0321 08:29:31.974687 4696 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 08:29:31 crc kubenswrapper[4696]: E0321 08:29:31.974751 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9bbfb96a-607b-47df-83df-7c75dd94ad71-metrics-certs podName:9bbfb96a-607b-47df-83df-7c75dd94ad71 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:33.974736486 +0000 UTC m=+108.095617199 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9bbfb96a-607b-47df-83df-7c75dd94ad71-metrics-certs") pod "network-metrics-daemon-w95sn" (UID: "9bbfb96a-607b-47df-83df-7c75dd94ad71") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 08:29:32 crc kubenswrapper[4696]: I0321 08:29:32.039643 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" event={"ID":"9f0e7707-cb43-4f15-b309-7f651da35ddf","Type":"ContainerStarted","Data":"2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a"} Mar 21 08:29:32 crc kubenswrapper[4696]: I0321 08:29:32.039682 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" event={"ID":"9f0e7707-cb43-4f15-b309-7f651da35ddf","Type":"ContainerStarted","Data":"cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7"} Mar 21 08:29:32 crc kubenswrapper[4696]: I0321 08:29:32.039691 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" event={"ID":"9f0e7707-cb43-4f15-b309-7f651da35ddf","Type":"ContainerStarted","Data":"92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59"} Mar 21 08:29:32 crc kubenswrapper[4696]: I0321 08:29:32.039699 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" event={"ID":"9f0e7707-cb43-4f15-b309-7f651da35ddf","Type":"ContainerStarted","Data":"42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054"} Mar 21 08:29:32 crc kubenswrapper[4696]: I0321 08:29:32.040732 4696 generic.go:334] "Generic (PLEG): container finished" podID="1472d11d-4da5-4423-99b1-4dd73e3e9b9b" containerID="8909ac6392af8ae740949680f83168611e7bfb38c64924b96abaa901d8c07880" exitCode=0 Mar 21 08:29:32 crc kubenswrapper[4696]: I0321 08:29:32.040894 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4rp52" event={"ID":"1472d11d-4da5-4423-99b1-4dd73e3e9b9b","Type":"ContainerDied","Data":"8909ac6392af8ae740949680f83168611e7bfb38c64924b96abaa901d8c07880"} Mar 21 08:29:32 crc kubenswrapper[4696]: I0321 08:29:32.063063 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-t6d75" podStartSLOduration=66.063043794 podStartE2EDuration="1m6.063043794s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:31.182197818 +0000 UTC m=+105.303078531" watchObservedRunningTime="2026-03-21 08:29:32.063043794 +0000 UTC m=+106.183924527" Mar 21 08:29:32 crc kubenswrapper[4696]: I0321 08:29:32.534376 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w95sn" Mar 21 08:29:32 crc kubenswrapper[4696]: E0321 08:29:32.534788 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w95sn" podUID="9bbfb96a-607b-47df-83df-7c75dd94ad71" Mar 21 08:29:33 crc kubenswrapper[4696]: I0321 08:29:33.046874 4696 generic.go:334] "Generic (PLEG): container finished" podID="1472d11d-4da5-4423-99b1-4dd73e3e9b9b" containerID="d51d5b76e707adeef476481f86c8b176fc19e1f3baed451147a238c191136817" exitCode=0 Mar 21 08:29:33 crc kubenswrapper[4696]: I0321 08:29:33.046994 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4rp52" event={"ID":"1472d11d-4da5-4423-99b1-4dd73e3e9b9b","Type":"ContainerDied","Data":"d51d5b76e707adeef476481f86c8b176fc19e1f3baed451147a238c191136817"} Mar 21 08:29:33 crc kubenswrapper[4696]: I0321 08:29:33.054419 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" event={"ID":"9f0e7707-cb43-4f15-b309-7f651da35ddf","Type":"ContainerStarted","Data":"4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8"} Mar 21 08:29:33 crc kubenswrapper[4696]: I0321 08:29:33.054720 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" event={"ID":"9f0e7707-cb43-4f15-b309-7f651da35ddf","Type":"ContainerStarted","Data":"beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a"} Mar 21 08:29:33 crc kubenswrapper[4696]: I0321 08:29:33.534460 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:33 crc kubenswrapper[4696]: I0321 08:29:33.534531 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:33 crc kubenswrapper[4696]: I0321 08:29:33.534586 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:33 crc kubenswrapper[4696]: E0321 08:29:33.534601 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 08:29:33 crc kubenswrapper[4696]: E0321 08:29:33.534743 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 08:29:33 crc kubenswrapper[4696]: E0321 08:29:33.534974 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 08:29:33 crc kubenswrapper[4696]: I0321 08:29:33.999036 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9bbfb96a-607b-47df-83df-7c75dd94ad71-metrics-certs\") pod \"network-metrics-daemon-w95sn\" (UID: \"9bbfb96a-607b-47df-83df-7c75dd94ad71\") " pod="openshift-multus/network-metrics-daemon-w95sn" Mar 21 08:29:33 crc kubenswrapper[4696]: E0321 08:29:33.999195 4696 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 08:29:33 crc kubenswrapper[4696]: E0321 08:29:33.999280 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9bbfb96a-607b-47df-83df-7c75dd94ad71-metrics-certs podName:9bbfb96a-607b-47df-83df-7c75dd94ad71 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:37.999255182 +0000 UTC m=+112.120135915 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9bbfb96a-607b-47df-83df-7c75dd94ad71-metrics-certs") pod "network-metrics-daemon-w95sn" (UID: "9bbfb96a-607b-47df-83df-7c75dd94ad71") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 08:29:34 crc kubenswrapper[4696]: I0321 08:29:34.059414 4696 generic.go:334] "Generic (PLEG): container finished" podID="1472d11d-4da5-4423-99b1-4dd73e3e9b9b" containerID="1717ab299765b278809bd859ff03558bb931f409324a17cc4394e1dd73adc81b" exitCode=0 Mar 21 08:29:34 crc kubenswrapper[4696]: I0321 08:29:34.059469 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4rp52" event={"ID":"1472d11d-4da5-4423-99b1-4dd73e3e9b9b","Type":"ContainerDied","Data":"1717ab299765b278809bd859ff03558bb931f409324a17cc4394e1dd73adc81b"} Mar 21 08:29:34 crc kubenswrapper[4696]: I0321 08:29:34.534491 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w95sn" Mar 21 08:29:34 crc kubenswrapper[4696]: E0321 08:29:34.534609 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w95sn" podUID="9bbfb96a-607b-47df-83df-7c75dd94ad71" Mar 21 08:29:35 crc kubenswrapper[4696]: I0321 08:29:35.066115 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" event={"ID":"9f0e7707-cb43-4f15-b309-7f651da35ddf","Type":"ContainerStarted","Data":"7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0"} Mar 21 08:29:35 crc kubenswrapper[4696]: I0321 08:29:35.068866 4696 generic.go:334] "Generic (PLEG): container finished" podID="1472d11d-4da5-4423-99b1-4dd73e3e9b9b" containerID="924084b0c28f404cc8b50827c98c6ae1c875e6fe533d6a2dcd184292968eaf7b" exitCode=0 Mar 21 08:29:35 crc kubenswrapper[4696]: I0321 08:29:35.068926 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4rp52" event={"ID":"1472d11d-4da5-4423-99b1-4dd73e3e9b9b","Type":"ContainerDied","Data":"924084b0c28f404cc8b50827c98c6ae1c875e6fe533d6a2dcd184292968eaf7b"} Mar 21 08:29:35 crc kubenswrapper[4696]: I0321 08:29:35.534430 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:35 crc kubenswrapper[4696]: E0321 08:29:35.534567 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 08:29:35 crc kubenswrapper[4696]: I0321 08:29:35.534572 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:35 crc kubenswrapper[4696]: I0321 08:29:35.534630 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:35 crc kubenswrapper[4696]: E0321 08:29:35.534665 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 08:29:35 crc kubenswrapper[4696]: E0321 08:29:35.534852 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 08:29:36 crc kubenswrapper[4696]: I0321 08:29:36.076547 4696 generic.go:334] "Generic (PLEG): container finished" podID="1472d11d-4da5-4423-99b1-4dd73e3e9b9b" containerID="4450acb358ce6032bb7d67d3761d353ea06441daf54279d70c205d43448845b0" exitCode=0 Mar 21 08:29:36 crc kubenswrapper[4696]: I0321 08:29:36.076603 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4rp52" event={"ID":"1472d11d-4da5-4423-99b1-4dd73e3e9b9b","Type":"ContainerDied","Data":"4450acb358ce6032bb7d67d3761d353ea06441daf54279d70c205d43448845b0"} Mar 21 08:29:36 crc kubenswrapper[4696]: I0321 08:29:36.533773 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w95sn" Mar 21 08:29:36 crc kubenswrapper[4696]: E0321 08:29:36.536756 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w95sn" podUID="9bbfb96a-607b-47df-83df-7c75dd94ad71" Mar 21 08:29:37 crc kubenswrapper[4696]: I0321 08:29:37.087289 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4rp52" event={"ID":"1472d11d-4da5-4423-99b1-4dd73e3e9b9b","Type":"ContainerStarted","Data":"e9436e9ea0c9572c471a3c13dc7648f925c0a8d9ff170da8f91e8ef0dccb7f77"} Mar 21 08:29:37 crc kubenswrapper[4696]: I0321 08:29:37.095143 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" event={"ID":"9f0e7707-cb43-4f15-b309-7f651da35ddf","Type":"ContainerStarted","Data":"4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281"} Mar 21 08:29:37 crc kubenswrapper[4696]: I0321 08:29:37.095757 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:37 crc kubenswrapper[4696]: I0321 08:29:37.124707 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-4rp52" podStartSLOduration=71.124666919 podStartE2EDuration="1m11.124666919s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:37.121628059 +0000 UTC m=+111.242508772" watchObservedRunningTime="2026-03-21 08:29:37.124666919 +0000 UTC m=+111.245547732" Mar 21 08:29:37 crc kubenswrapper[4696]: I0321 08:29:37.139798 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:37 crc kubenswrapper[4696]: I0321 08:29:37.208381 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" podStartSLOduration=71.208352005 podStartE2EDuration="1m11.208352005s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:37.158071907 +0000 UTC m=+111.278952640" watchObservedRunningTime="2026-03-21 08:29:37.208352005 +0000 UTC m=+111.329232728" Mar 21 08:29:37 crc kubenswrapper[4696]: I0321 08:29:37.534549 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:37 crc kubenswrapper[4696]: I0321 08:29:37.534613 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:37 crc kubenswrapper[4696]: E0321 08:29:37.534693 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 08:29:37 crc kubenswrapper[4696]: I0321 08:29:37.534783 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:37 crc kubenswrapper[4696]: E0321 08:29:37.534981 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 08:29:37 crc kubenswrapper[4696]: E0321 08:29:37.535034 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 08:29:38 crc kubenswrapper[4696]: I0321 08:29:38.046457 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9bbfb96a-607b-47df-83df-7c75dd94ad71-metrics-certs\") pod \"network-metrics-daemon-w95sn\" (UID: \"9bbfb96a-607b-47df-83df-7c75dd94ad71\") " pod="openshift-multus/network-metrics-daemon-w95sn" Mar 21 08:29:38 crc kubenswrapper[4696]: E0321 08:29:38.046627 4696 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 08:29:38 crc kubenswrapper[4696]: E0321 08:29:38.046689 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9bbfb96a-607b-47df-83df-7c75dd94ad71-metrics-certs podName:9bbfb96a-607b-47df-83df-7c75dd94ad71 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:46.046673954 +0000 UTC m=+120.167554667 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9bbfb96a-607b-47df-83df-7c75dd94ad71-metrics-certs") pod "network-metrics-daemon-w95sn" (UID: "9bbfb96a-607b-47df-83df-7c75dd94ad71") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 08:29:38 crc kubenswrapper[4696]: I0321 08:29:38.099109 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:38 crc kubenswrapper[4696]: I0321 08:29:38.099148 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:38 crc kubenswrapper[4696]: I0321 08:29:38.118757 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:29:38 crc kubenswrapper[4696]: I0321 08:29:38.534649 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w95sn" Mar 21 08:29:38 crc kubenswrapper[4696]: E0321 08:29:38.534953 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w95sn" podUID="9bbfb96a-607b-47df-83df-7c75dd94ad71" Mar 21 08:29:39 crc kubenswrapper[4696]: I0321 08:29:39.014610 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-w95sn"] Mar 21 08:29:39 crc kubenswrapper[4696]: I0321 08:29:39.101617 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w95sn" Mar 21 08:29:39 crc kubenswrapper[4696]: E0321 08:29:39.102122 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w95sn" podUID="9bbfb96a-607b-47df-83df-7c75dd94ad71" Mar 21 08:29:39 crc kubenswrapper[4696]: I0321 08:29:39.533999 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:39 crc kubenswrapper[4696]: I0321 08:29:39.534020 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:39 crc kubenswrapper[4696]: I0321 08:29:39.534108 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:39 crc kubenswrapper[4696]: E0321 08:29:39.534268 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 21 08:29:39 crc kubenswrapper[4696]: E0321 08:29:39.534313 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 21 08:29:39 crc kubenswrapper[4696]: E0321 08:29:39.534380 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 21 08:29:40 crc kubenswrapper[4696]: I0321 08:29:40.533894 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w95sn" Mar 21 08:29:40 crc kubenswrapper[4696]: E0321 08:29:40.534001 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-w95sn" podUID="9bbfb96a-607b-47df-83df-7c75dd94ad71" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.221807 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.464632 4696 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.464935 4696 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.513767 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ptg5t"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.514431 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.518760 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.522612 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rvwvm"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.523184 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-f5n2g"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.523609 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.523751 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.525081 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.525389 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-xt5jj"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.526438 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xt5jj" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.527253 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-xblff"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.527919 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.528583 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.529463 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.530081 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vnbgm"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.530546 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vnbgm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.535046 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.535070 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.535195 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.535338 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.535494 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.536543 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-9thjt"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.537085 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-9thjt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.537166 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-c9v5l"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.537557 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-c9v5l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.544053 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.544262 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.544401 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.544538 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.545483 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.545586 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.545645 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.550191 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.550313 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.550643 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.550673 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.550872 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.550937 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.551116 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.550221 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.551437 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.551916 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.552060 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.552190 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-khvnh"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.552610 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.552776 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qc5tq"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.553006 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.553120 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.553190 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.553235 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5thcd"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.553120 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.553369 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.553416 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-khvnh" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.553065 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.553486 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.553591 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5thcd" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.553591 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qc5tq" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.557178 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.557543 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.557567 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.557736 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.559785 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.560045 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.560211 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.560403 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.560587 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.560750 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.566529 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.570698 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.571193 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.571216 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.582548 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bp44p"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.584870 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.585257 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-bp44p" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.600739 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.600810 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.601184 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.601196 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.601492 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.601578 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e24de32f-7778-4fac-815b-fa4d2a44dd2f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.601621 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-serving-cert\") pod \"controller-manager-879f6c89f-ptg5t\" (UID: \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.601644 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/fb015b47-dbed-4cc5-a8d4-ab5d55eb1514-etcd-ca\") pod \"etcd-operator-b45778765-rvwvm\" (UID: \"fb015b47-dbed-4cc5-a8d4-ab5d55eb1514\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.601659 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e24de32f-7778-4fac-815b-fa4d2a44dd2f-encryption-config\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.601680 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc6gp\" (UniqueName: \"kubernetes.io/projected/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-kube-api-access-lc6gp\") pod \"route-controller-manager-6576b87f9c-xnns6\" (UID: \"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.601700 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e24de32f-7778-4fac-815b-fa4d2a44dd2f-config\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.601716 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-serving-cert\") pod \"route-controller-manager-6576b87f9c-xnns6\" (UID: \"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.601731 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-service-ca\") pod \"console-f9d7485db-f5n2g\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.601758 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e24de32f-7778-4fac-815b-fa4d2a44dd2f-node-pullsecrets\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.601775 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-config\") pod \"route-controller-manager-6576b87f9c-xnns6\" (UID: \"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.601842 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e24de32f-7778-4fac-815b-fa4d2a44dd2f-etcd-client\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.601876 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e24de32f-7778-4fac-815b-fa4d2a44dd2f-serving-cert\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.601922 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e24de32f-7778-4fac-815b-fa4d2a44dd2f-image-import-ca\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.601951 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-config\") pod \"controller-manager-879f6c89f-ptg5t\" (UID: \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602014 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78c4d6cc-8ad5-4d4b-b20d-bebe8259e18e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vnbgm\" (UID: \"78c4d6cc-8ad5-4d4b-b20d-bebe8259e18e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vnbgm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602086 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9q56\" (UniqueName: \"kubernetes.io/projected/d1d05966-d105-4c45-bfa2-e66a67882baf-kube-api-access-x9q56\") pod \"console-f9d7485db-f5n2g\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602110 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68a1b8db-aec4-4681-b1f0-87761e34c518-serving-cert\") pod \"openshift-config-operator-7777fb866f-xt5jj\" (UID: \"68a1b8db-aec4-4681-b1f0-87761e34c518\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xt5jj" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602132 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ptg5t\" (UID: \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602155 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-console-config\") pod \"console-f9d7485db-f5n2g\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602174 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvhrn\" (UniqueName: \"kubernetes.io/projected/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-kube-api-access-qvhrn\") pod \"controller-manager-879f6c89f-ptg5t\" (UID: \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602199 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-trusted-ca-bundle\") pod \"console-f9d7485db-f5n2g\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602218 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e24de32f-7778-4fac-815b-fa4d2a44dd2f-audit-dir\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602250 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fb015b47-dbed-4cc5-a8d4-ab5d55eb1514-etcd-client\") pod \"etcd-operator-b45778765-rvwvm\" (UID: \"fb015b47-dbed-4cc5-a8d4-ab5d55eb1514\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602269 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-oauth-serving-cert\") pod \"console-f9d7485db-f5n2g\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602296 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d1d05966-d105-4c45-bfa2-e66a67882baf-console-oauth-config\") pod \"console-f9d7485db-f5n2g\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602321 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e24de32f-7778-4fac-815b-fa4d2a44dd2f-audit\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602345 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d1d05966-d105-4c45-bfa2-e66a67882baf-console-serving-cert\") pod \"console-f9d7485db-f5n2g\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602365 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e24de32f-7778-4fac-815b-fa4d2a44dd2f-etcd-serving-ca\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602387 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/68a1b8db-aec4-4681-b1f0-87761e34c518-available-featuregates\") pod \"openshift-config-operator-7777fb866f-xt5jj\" (UID: \"68a1b8db-aec4-4681-b1f0-87761e34c518\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xt5jj" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602499 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6v76\" (UniqueName: \"kubernetes.io/projected/e24de32f-7778-4fac-815b-fa4d2a44dd2f-kube-api-access-v6v76\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602551 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78c4d6cc-8ad5-4d4b-b20d-bebe8259e18e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vnbgm\" (UID: \"78c4d6cc-8ad5-4d4b-b20d-bebe8259e18e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vnbgm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602594 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-client-ca\") pod \"route-controller-manager-6576b87f9c-xnns6\" (UID: \"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602628 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb015b47-dbed-4cc5-a8d4-ab5d55eb1514-config\") pod \"etcd-operator-b45778765-rvwvm\" (UID: \"fb015b47-dbed-4cc5-a8d4-ab5d55eb1514\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602661 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nspjm\" (UniqueName: \"kubernetes.io/projected/68a1b8db-aec4-4681-b1f0-87761e34c518-kube-api-access-nspjm\") pod \"openshift-config-operator-7777fb866f-xt5jj\" (UID: \"68a1b8db-aec4-4681-b1f0-87761e34c518\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xt5jj" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602700 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48rwn\" (UniqueName: \"kubernetes.io/projected/78c4d6cc-8ad5-4d4b-b20d-bebe8259e18e-kube-api-access-48rwn\") pod \"openshift-controller-manager-operator-756b6f6bc6-vnbgm\" (UID: \"78c4d6cc-8ad5-4d4b-b20d-bebe8259e18e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vnbgm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602746 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr74r\" (UniqueName: \"kubernetes.io/projected/fb015b47-dbed-4cc5-a8d4-ab5d55eb1514-kube-api-access-lr74r\") pod \"etcd-operator-b45778765-rvwvm\" (UID: \"fb015b47-dbed-4cc5-a8d4-ab5d55eb1514\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602803 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602803 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb015b47-dbed-4cc5-a8d4-ab5d55eb1514-serving-cert\") pod \"etcd-operator-b45778765-rvwvm\" (UID: \"fb015b47-dbed-4cc5-a8d4-ab5d55eb1514\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602958 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/fb015b47-dbed-4cc5-a8d4-ab5d55eb1514-etcd-service-ca\") pod \"etcd-operator-b45778765-rvwvm\" (UID: \"fb015b47-dbed-4cc5-a8d4-ab5d55eb1514\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.602991 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-client-ca\") pod \"controller-manager-879f6c89f-ptg5t\" (UID: \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.603186 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.603322 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.603415 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.603504 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.606011 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.606128 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.611539 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.611724 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.612126 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.612496 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.612744 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.613063 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.614724 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.614771 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.614951 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.614951 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.615121 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.615199 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qvm82"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.615233 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.615371 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.615925 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.615982 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-h5tfm"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.616125 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.616327 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.616530 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-h5tfm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.617722 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-6m4ch"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.620891 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.614595 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.624793 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.625281 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.625492 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.627300 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.629536 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-jdfkn"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.627327 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.627528 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.627611 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.627645 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.627708 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.627941 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.627978 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.628029 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.628063 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.628883 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.628915 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.628944 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.628973 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.629012 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.634480 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.635978 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.636087 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.636267 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.636383 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.636397 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.629622 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-6m4ch" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.670551 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.671089 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ptg5t"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.671111 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vnbgm"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.671122 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rvwvm"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.671132 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-f5n2g"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.671142 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-xt5jj"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.671152 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-xblff"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.671162 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd5f5"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.671446 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.671533 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd5f5" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.671920 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-jdfkn" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.672051 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.675731 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2q58n"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.676165 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.676268 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2q58n" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.677679 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.678200 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.678911 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-lxfpb"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.679676 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c46nr"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.680057 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-59mfl"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.680482 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.680521 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfpb" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.680547 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qs7dj"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.680700 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.680490 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-59mfl" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.681255 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qs7dj" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.681961 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-sdgwl"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.683340 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.687175 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.687183 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.687959 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.688792 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-sdgwl" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.689323 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5mlz9"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.689679 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5mlz9" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.689933 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ntmrq"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.690537 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ntmrq" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.691349 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-9thjt"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.692512 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r656w"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.692904 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5vrbx"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.693221 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.693282 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r656w" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.693481 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-bbpqw"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.693919 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bbpqw" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.694964 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.696789 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.698446 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bp44p"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.700935 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qvm82"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.700970 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-jdfkn"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.703456 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-6m4ch"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.704842 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f2a60c2e-1edd-4c96-b263-e84dec7ed4a2-service-ca-bundle\") pod \"authentication-operator-69f744f599-bp44p\" (UID: \"f2a60c2e-1edd-4c96-b263-e84dec7ed4a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bp44p" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.704871 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.704895 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ad264986-2840-4a2a-bee9-9ea192cdbe86-machine-approver-tls\") pod \"machine-approver-56656f9798-khvnh\" (UID: \"ad264986-2840-4a2a-bee9-9ea192cdbe86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-khvnh" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.704913 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.704933 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9q56\" (UniqueName: \"kubernetes.io/projected/d1d05966-d105-4c45-bfa2-e66a67882baf-kube-api-access-x9q56\") pod \"console-f9d7485db-f5n2g\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.704951 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68a1b8db-aec4-4681-b1f0-87761e34c518-serving-cert\") pod \"openshift-config-operator-7777fb866f-xt5jj\" (UID: \"68a1b8db-aec4-4681-b1f0-87761e34c518\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xt5jj" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.704966 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcd8k\" (UniqueName: \"kubernetes.io/projected/13b3ee79-f56b-4fff-ad26-ac72a3ef5017-kube-api-access-kcd8k\") pod \"cluster-image-registry-operator-dc59b4c8b-c9v5l\" (UID: \"13b3ee79-f56b-4fff-ad26-ac72a3ef5017\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-c9v5l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.704983 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/851ce288-d60d-4875-ae01-cc13268fec8c-audit-dir\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.704999 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ptg5t\" (UID: \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.705014 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e26909f3-14e6-4f5a-9ea9-5b44ac3aa56a-metrics-tls\") pod \"dns-operator-744455d44c-6m4ch\" (UID: \"e26909f3-14e6-4f5a-9ea9-5b44ac3aa56a\") " pod="openshift-dns-operator/dns-operator-744455d44c-6m4ch" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.705031 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/13b3ee79-f56b-4fff-ad26-ac72a3ef5017-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-c9v5l\" (UID: \"13b3ee79-f56b-4fff-ad26-ac72a3ef5017\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-c9v5l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.705048 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-console-config\") pod \"console-f9d7485db-f5n2g\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.705065 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvhrn\" (UniqueName: \"kubernetes.io/projected/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-kube-api-access-qvhrn\") pod \"controller-manager-879f6c89f-ptg5t\" (UID: \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.705080 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3239cfe6-5d28-4efc-b0b3-53699ca2a449-config\") pod \"console-operator-58897d9998-9thjt\" (UID: \"3239cfe6-5d28-4efc-b0b3-53699ca2a449\") " pod="openshift-console-operator/console-operator-58897d9998-9thjt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.705096 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-trusted-ca-bundle\") pod \"console-f9d7485db-f5n2g\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.705112 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e24de32f-7778-4fac-815b-fa4d2a44dd2f-audit-dir\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.705136 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fb015b47-dbed-4cc5-a8d4-ab5d55eb1514-etcd-client\") pod \"etcd-operator-b45778765-rvwvm\" (UID: \"fb015b47-dbed-4cc5-a8d4-ab5d55eb1514\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.705409 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e24de32f-7778-4fac-815b-fa4d2a44dd2f-audit-dir\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.706118 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-console-config\") pod \"console-f9d7485db-f5n2g\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.706445 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-oauth-serving-cert\") pod \"console-f9d7485db-f5n2g\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.706506 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.706533 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d1d05966-d105-4c45-bfa2-e66a67882baf-console-oauth-config\") pod \"console-f9d7485db-f5n2g\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.706552 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwzfz\" (UniqueName: \"kubernetes.io/projected/f2a60c2e-1edd-4c96-b263-e84dec7ed4a2-kube-api-access-gwzfz\") pod \"authentication-operator-69f744f599-bp44p\" (UID: \"f2a60c2e-1edd-4c96-b263-e84dec7ed4a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bp44p" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.706580 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-trusted-ca-bundle\") pod \"console-f9d7485db-f5n2g\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.707262 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ptg5t\" (UID: \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.707419 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-oauth-serving-cert\") pod \"console-f9d7485db-f5n2g\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708196 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708263 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e24de32f-7778-4fac-815b-fa4d2a44dd2f-audit\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708313 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85rms\" (UniqueName: \"kubernetes.io/projected/3239cfe6-5d28-4efc-b0b3-53699ca2a449-kube-api-access-85rms\") pod \"console-operator-58897d9998-9thjt\" (UID: \"3239cfe6-5d28-4efc-b0b3-53699ca2a449\") " pod="openshift-console-operator/console-operator-58897d9998-9thjt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708343 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708399 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d1d05966-d105-4c45-bfa2-e66a67882baf-console-serving-cert\") pod \"console-f9d7485db-f5n2g\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708426 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e24de32f-7778-4fac-815b-fa4d2a44dd2f-etcd-serving-ca\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708472 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/68a1b8db-aec4-4681-b1f0-87761e34c518-available-featuregates\") pod \"openshift-config-operator-7777fb866f-xt5jj\" (UID: \"68a1b8db-aec4-4681-b1f0-87761e34c518\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xt5jj" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708503 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6v76\" (UniqueName: \"kubernetes.io/projected/e24de32f-7778-4fac-815b-fa4d2a44dd2f-kube-api-access-v6v76\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708521 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78c4d6cc-8ad5-4d4b-b20d-bebe8259e18e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vnbgm\" (UID: \"78c4d6cc-8ad5-4d4b-b20d-bebe8259e18e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vnbgm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708584 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f62330e9-cdc5-469b-82ba-3364ce173686-config\") pod \"machine-api-operator-5694c8668f-h5tfm\" (UID: \"f62330e9-cdc5-469b-82ba-3364ce173686\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h5tfm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708609 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-client-ca\") pod \"route-controller-manager-6576b87f9c-xnns6\" (UID: \"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708651 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb015b47-dbed-4cc5-a8d4-ab5d55eb1514-config\") pod \"etcd-operator-b45778765-rvwvm\" (UID: \"fb015b47-dbed-4cc5-a8d4-ab5d55eb1514\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708669 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nspjm\" (UniqueName: \"kubernetes.io/projected/68a1b8db-aec4-4681-b1f0-87761e34c518-kube-api-access-nspjm\") pod \"openshift-config-operator-7777fb866f-xt5jj\" (UID: \"68a1b8db-aec4-4681-b1f0-87761e34c518\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xt5jj" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708690 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/dc140eb2-a579-45bc-9fb4-247c254533ea-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qc5tq\" (UID: \"dc140eb2-a579-45bc-9fb4-247c254533ea\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qc5tq" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708751 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f62330e9-cdc5-469b-82ba-3364ce173686-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-h5tfm\" (UID: \"f62330e9-cdc5-469b-82ba-3364ce173686\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h5tfm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708777 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48rwn\" (UniqueName: \"kubernetes.io/projected/78c4d6cc-8ad5-4d4b-b20d-bebe8259e18e-kube-api-access-48rwn\") pod \"openshift-controller-manager-operator-756b6f6bc6-vnbgm\" (UID: \"78c4d6cc-8ad5-4d4b-b20d-bebe8259e18e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vnbgm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708825 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708843 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6fqx\" (UniqueName: \"kubernetes.io/projected/851ce288-d60d-4875-ae01-cc13268fec8c-kube-api-access-x6fqx\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708863 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr74r\" (UniqueName: \"kubernetes.io/projected/fb015b47-dbed-4cc5-a8d4-ab5d55eb1514-kube-api-access-lr74r\") pod \"etcd-operator-b45778765-rvwvm\" (UID: \"fb015b47-dbed-4cc5-a8d4-ab5d55eb1514\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708901 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f2a60c2e-1edd-4c96-b263-e84dec7ed4a2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bp44p\" (UID: \"f2a60c2e-1edd-4c96-b263-e84dec7ed4a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bp44p" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708921 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf8bw\" (UniqueName: \"kubernetes.io/projected/f62330e9-cdc5-469b-82ba-3364ce173686-kube-api-access-rf8bw\") pod \"machine-api-operator-5694c8668f-h5tfm\" (UID: \"f62330e9-cdc5-469b-82ba-3364ce173686\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h5tfm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.708938 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ad264986-2840-4a2a-bee9-9ea192cdbe86-auth-proxy-config\") pod \"machine-approver-56656f9798-khvnh\" (UID: \"ad264986-2840-4a2a-bee9-9ea192cdbe86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-khvnh" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.709002 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb015b47-dbed-4cc5-a8d4-ab5d55eb1514-serving-cert\") pod \"etcd-operator-b45778765-rvwvm\" (UID: \"fb015b47-dbed-4cc5-a8d4-ab5d55eb1514\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.709049 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/fb015b47-dbed-4cc5-a8d4-ab5d55eb1514-etcd-service-ca\") pod \"etcd-operator-b45778765-rvwvm\" (UID: \"fb015b47-dbed-4cc5-a8d4-ab5d55eb1514\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.709073 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-client-ca\") pod \"controller-manager-879f6c89f-ptg5t\" (UID: \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.709091 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b83c2afe-6390-4518-a5c1-c6ca14ea4629-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5thcd\" (UID: \"b83c2afe-6390-4518-a5c1-c6ca14ea4629\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5thcd" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.709258 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3239cfe6-5d28-4efc-b0b3-53699ca2a449-trusted-ca\") pod \"console-operator-58897d9998-9thjt\" (UID: \"3239cfe6-5d28-4efc-b0b3-53699ca2a449\") " pod="openshift-console-operator/console-operator-58897d9998-9thjt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.709320 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e24de32f-7778-4fac-815b-fa4d2a44dd2f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.709340 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-serving-cert\") pod \"controller-manager-879f6c89f-ptg5t\" (UID: \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.709366 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/fb015b47-dbed-4cc5-a8d4-ab5d55eb1514-etcd-ca\") pod \"etcd-operator-b45778765-rvwvm\" (UID: \"fb015b47-dbed-4cc5-a8d4-ab5d55eb1514\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.709409 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.709429 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e24de32f-7778-4fac-815b-fa4d2a44dd2f-encryption-config\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.709465 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2a60c2e-1edd-4c96-b263-e84dec7ed4a2-config\") pod \"authentication-operator-69f744f599-bp44p\" (UID: \"f2a60c2e-1edd-4c96-b263-e84dec7ed4a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bp44p" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.709493 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2a60c2e-1edd-4c96-b263-e84dec7ed4a2-serving-cert\") pod \"authentication-operator-69f744f599-bp44p\" (UID: \"f2a60c2e-1edd-4c96-b263-e84dec7ed4a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bp44p" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.718988 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68a1b8db-aec4-4681-b1f0-87761e34c518-serving-cert\") pod \"openshift-config-operator-7777fb866f-xt5jj\" (UID: \"68a1b8db-aec4-4681-b1f0-87761e34c518\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xt5jj" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.720116 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e24de32f-7778-4fac-815b-fa4d2a44dd2f-audit\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.720550 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fb015b47-dbed-4cc5-a8d4-ab5d55eb1514-etcd-client\") pod \"etcd-operator-b45778765-rvwvm\" (UID: \"fb015b47-dbed-4cc5-a8d4-ab5d55eb1514\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.720588 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-h5tfm"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.721992 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d1d05966-d105-4c45-bfa2-e66a67882baf-console-oauth-config\") pod \"console-f9d7485db-f5n2g\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.722969 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/fb015b47-dbed-4cc5-a8d4-ab5d55eb1514-etcd-service-ca\") pod \"etcd-operator-b45778765-rvwvm\" (UID: \"fb015b47-dbed-4cc5-a8d4-ab5d55eb1514\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.723010 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-client-ca\") pod \"controller-manager-879f6c89f-ptg5t\" (UID: \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.724230 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.725206 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e24de32f-7778-4fac-815b-fa4d2a44dd2f-etcd-serving-ca\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.726650 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-client-ca\") pod \"route-controller-manager-6576b87f9c-xnns6\" (UID: \"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.727225 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-serving-cert\") pod \"controller-manager-879f6c89f-ptg5t\" (UID: \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.727346 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e24de32f-7778-4fac-815b-fa4d2a44dd2f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.727384 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb015b47-dbed-4cc5-a8d4-ab5d55eb1514-config\") pod \"etcd-operator-b45778765-rvwvm\" (UID: \"fb015b47-dbed-4cc5-a8d4-ab5d55eb1514\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.728360 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/fb015b47-dbed-4cc5-a8d4-ab5d55eb1514-etcd-ca\") pod \"etcd-operator-b45778765-rvwvm\" (UID: \"fb015b47-dbed-4cc5-a8d4-ab5d55eb1514\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.728380 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f62330e9-cdc5-469b-82ba-3364ce173686-images\") pod \"machine-api-operator-5694c8668f-h5tfm\" (UID: \"f62330e9-cdc5-469b-82ba-3364ce173686\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h5tfm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.729336 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.729468 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc6gp\" (UniqueName: \"kubernetes.io/projected/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-kube-api-access-lc6gp\") pod \"route-controller-manager-6576b87f9c-xnns6\" (UID: \"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.729522 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/13b3ee79-f56b-4fff-ad26-ac72a3ef5017-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-c9v5l\" (UID: \"13b3ee79-f56b-4fff-ad26-ac72a3ef5017\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-c9v5l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.729667 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvvhj\" (UniqueName: \"kubernetes.io/projected/ad264986-2840-4a2a-bee9-9ea192cdbe86-kube-api-access-zvvhj\") pod \"machine-approver-56656f9798-khvnh\" (UID: \"ad264986-2840-4a2a-bee9-9ea192cdbe86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-khvnh" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.729712 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b83c2afe-6390-4518-a5c1-c6ca14ea4629-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5thcd\" (UID: \"b83c2afe-6390-4518-a5c1-c6ca14ea4629\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5thcd" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.729765 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtzxg\" (UniqueName: \"kubernetes.io/projected/b83c2afe-6390-4518-a5c1-c6ca14ea4629-kube-api-access-qtzxg\") pod \"openshift-apiserver-operator-796bbdcf4f-5thcd\" (UID: \"b83c2afe-6390-4518-a5c1-c6ca14ea4629\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5thcd" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.730018 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e24de32f-7778-4fac-815b-fa4d2a44dd2f-config\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.730043 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/13b3ee79-f56b-4fff-ad26-ac72a3ef5017-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-c9v5l\" (UID: \"13b3ee79-f56b-4fff-ad26-ac72a3ef5017\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-c9v5l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.730091 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-serving-cert\") pod \"route-controller-manager-6576b87f9c-xnns6\" (UID: \"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.730116 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-service-ca\") pod \"console-f9d7485db-f5n2g\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.730139 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j69wn\" (UniqueName: \"kubernetes.io/projected/e26909f3-14e6-4f5a-9ea9-5b44ac3aa56a-kube-api-access-j69wn\") pod \"dns-operator-744455d44c-6m4ch\" (UID: \"e26909f3-14e6-4f5a-9ea9-5b44ac3aa56a\") " pod="openshift-dns-operator/dns-operator-744455d44c-6m4ch" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.730185 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.730214 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/68a1b8db-aec4-4681-b1f0-87761e34c518-available-featuregates\") pod \"openshift-config-operator-7777fb866f-xt5jj\" (UID: \"68a1b8db-aec4-4681-b1f0-87761e34c518\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xt5jj" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.730288 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e24de32f-7778-4fac-815b-fa4d2a44dd2f-node-pullsecrets\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.730351 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e24de32f-7778-4fac-815b-fa4d2a44dd2f-node-pullsecrets\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.730402 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-config\") pod \"route-controller-manager-6576b87f9c-xnns6\" (UID: \"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.730549 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e24de32f-7778-4fac-815b-fa4d2a44dd2f-etcd-client\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.730575 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e24de32f-7778-4fac-815b-fa4d2a44dd2f-serving-cert\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.730600 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3239cfe6-5d28-4efc-b0b3-53699ca2a449-serving-cert\") pod \"console-operator-58897d9998-9thjt\" (UID: \"3239cfe6-5d28-4efc-b0b3-53699ca2a449\") " pod="openshift-console-operator/console-operator-58897d9998-9thjt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.730625 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e24de32f-7778-4fac-815b-fa4d2a44dd2f-image-import-ca\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.730642 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-config\") pod \"controller-manager-879f6c89f-ptg5t\" (UID: \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.730666 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.730729 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-audit-policies\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.730753 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.730849 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78c4d6cc-8ad5-4d4b-b20d-bebe8259e18e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vnbgm\" (UID: \"78c4d6cc-8ad5-4d4b-b20d-bebe8259e18e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vnbgm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.730893 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r74wp\" (UniqueName: \"kubernetes.io/projected/dc140eb2-a579-45bc-9fb4-247c254533ea-kube-api-access-r74wp\") pod \"cluster-samples-operator-665b6dd947-qc5tq\" (UID: \"dc140eb2-a579-45bc-9fb4-247c254533ea\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qc5tq" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.730917 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad264986-2840-4a2a-bee9-9ea192cdbe86-config\") pod \"machine-approver-56656f9798-khvnh\" (UID: \"ad264986-2840-4a2a-bee9-9ea192cdbe86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-khvnh" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.731057 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e24de32f-7778-4fac-815b-fa4d2a44dd2f-config\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.732663 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-service-ca\") pod \"console-f9d7485db-f5n2g\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.733304 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-config\") pod \"route-controller-manager-6576b87f9c-xnns6\" (UID: \"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.733337 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78c4d6cc-8ad5-4d4b-b20d-bebe8259e18e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vnbgm\" (UID: \"78c4d6cc-8ad5-4d4b-b20d-bebe8259e18e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vnbgm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.733925 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78c4d6cc-8ad5-4d4b-b20d-bebe8259e18e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vnbgm\" (UID: \"78c4d6cc-8ad5-4d4b-b20d-bebe8259e18e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vnbgm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.735383 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-config\") pod \"controller-manager-879f6c89f-ptg5t\" (UID: \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.735595 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e24de32f-7778-4fac-815b-fa4d2a44dd2f-etcd-client\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.736696 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e24de32f-7778-4fac-815b-fa4d2a44dd2f-image-import-ca\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.737094 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-serving-cert\") pod \"route-controller-manager-6576b87f9c-xnns6\" (UID: \"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.737410 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.737938 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e24de32f-7778-4fac-815b-fa4d2a44dd2f-encryption-config\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.743200 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5thcd"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.744204 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb015b47-dbed-4cc5-a8d4-ab5d55eb1514-serving-cert\") pod \"etcd-operator-b45778765-rvwvm\" (UID: \"fb015b47-dbed-4cc5-a8d4-ab5d55eb1514\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.745238 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d1d05966-d105-4c45-bfa2-e66a67882baf-console-serving-cert\") pod \"console-f9d7485db-f5n2g\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.745310 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qc5tq"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.746357 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e24de32f-7778-4fac-815b-fa4d2a44dd2f-serving-cert\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.749751 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-65h6c"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.764462 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.764556 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-c9v5l"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.764588 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-f8hcm"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.764836 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-65h6c" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.765007 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.765214 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8hcm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.765336 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9s5ck"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.765495 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.765683 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.765875 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9s5ck" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.766035 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-44xzn"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.766131 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.766779 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qm5pb"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.767024 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-44xzn" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.769173 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.769572 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd5f5"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.769593 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.769604 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2q58n"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.769674 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.769951 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qm5pb" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.770183 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-59mfl"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.772202 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-lxfpb"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.774209 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qs7dj"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.774248 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.774257 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c46nr"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.776998 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-65h6c"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.777023 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5mlz9"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.777034 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r656w"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.779736 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5vrbx"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.779761 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.779773 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qm5pb"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.781908 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-f8hcm"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.781972 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.789335 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-chtnc"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.790398 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.790421 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ntmrq"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.790501 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-chtnc" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.791344 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.791758 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wshd6"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.792754 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-bbpqw"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.792892 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-wshd6" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.793970 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-chtnc"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.795517 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9s5ck"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.795935 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wshd6"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.796983 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.797113 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-wkbtr"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.797853 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.800947 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-4tb7m"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.801982 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-4tb7m"] Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.802071 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-4tb7m" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.820222 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832215 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/934c62b8-c7a7-4018-b5f1-7954b2baca00-webhook-cert\") pod \"packageserver-d55dfcdfc-h7fjx\" (UID: \"934c62b8-c7a7-4018-b5f1-7954b2baca00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832247 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4928d31-a7c1-47f0-a726-712de834e034-config\") pod \"kube-controller-manager-operator-78b949d7b-sd5f5\" (UID: \"b4928d31-a7c1-47f0-a726-712de834e034\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd5f5" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832267 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh4qr\" (UniqueName: \"kubernetes.io/projected/934c62b8-c7a7-4018-b5f1-7954b2baca00-kube-api-access-mh4qr\") pod \"packageserver-d55dfcdfc-h7fjx\" (UID: \"934c62b8-c7a7-4018-b5f1-7954b2baca00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832282 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hbtq\" (UniqueName: \"kubernetes.io/projected/6ef60c27-2062-4bc8-8c37-e13db97d29a3-kube-api-access-8hbtq\") pod \"collect-profiles-29568015-msx48\" (UID: \"6ef60c27-2062-4bc8-8c37-e13db97d29a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832298 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5b9d4194-45b6-445b-b882-1982630eaecf-proxy-tls\") pod \"machine-config-controller-84d6567774-bbpqw\" (UID: \"5b9d4194-45b6-445b-b882-1982630eaecf\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bbpqw" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832318 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3239cfe6-5d28-4efc-b0b3-53699ca2a449-serving-cert\") pod \"console-operator-58897d9998-9thjt\" (UID: \"3239cfe6-5d28-4efc-b0b3-53699ca2a449\") " pod="openshift-console-operator/console-operator-58897d9998-9thjt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832336 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832353 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ffb878ee-7cd7-4830-bb19-6b2a5e6bd1e0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-lxfpb\" (UID: \"ffb878ee-7cd7-4830-bb19-6b2a5e6bd1e0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfpb" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832404 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r74wp\" (UniqueName: \"kubernetes.io/projected/dc140eb2-a579-45bc-9fb4-247c254533ea-kube-api-access-r74wp\") pod \"cluster-samples-operator-665b6dd947-qc5tq\" (UID: \"dc140eb2-a579-45bc-9fb4-247c254533ea\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qc5tq" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832434 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad264986-2840-4a2a-bee9-9ea192cdbe86-config\") pod \"machine-approver-56656f9798-khvnh\" (UID: \"ad264986-2840-4a2a-bee9-9ea192cdbe86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-khvnh" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832454 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832476 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/44b39f4b-4213-452c-94f7-946308af1e2f-profile-collector-cert\") pod \"catalog-operator-68c6474976-h5n9r\" (UID: \"44b39f4b-4213-452c-94f7-946308af1e2f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832493 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ed334ee3-b168-4ac1-94a7-45aa902b9b32-auth-proxy-config\") pod \"machine-config-operator-74547568cd-4scwl\" (UID: \"ed334ee3-b168-4ac1-94a7-45aa902b9b32\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832526 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ad264986-2840-4a2a-bee9-9ea192cdbe86-machine-approver-tls\") pod \"machine-approver-56656f9798-khvnh\" (UID: \"ad264986-2840-4a2a-bee9-9ea192cdbe86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-khvnh" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832543 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/934c62b8-c7a7-4018-b5f1-7954b2baca00-tmpfs\") pod \"packageserver-d55dfcdfc-h7fjx\" (UID: \"934c62b8-c7a7-4018-b5f1-7954b2baca00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832560 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/91299f09-692d-4c4e-a83c-22fcd785c1ef-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832577 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/13b3ee79-f56b-4fff-ad26-ac72a3ef5017-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-c9v5l\" (UID: \"13b3ee79-f56b-4fff-ad26-ac72a3ef5017\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-c9v5l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832592 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e26909f3-14e6-4f5a-9ea9-5b44ac3aa56a-metrics-tls\") pod \"dns-operator-744455d44c-6m4ch\" (UID: \"e26909f3-14e6-4f5a-9ea9-5b44ac3aa56a\") " pod="openshift-dns-operator/dns-operator-744455d44c-6m4ch" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832624 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14c95e94-4284-423b-b1cf-8df8ef663d65-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qs7dj\" (UID: \"14c95e94-4284-423b-b1cf-8df8ef663d65\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qs7dj" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832645 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnwn4\" (UniqueName: \"kubernetes.io/projected/5b9d4194-45b6-445b-b882-1982630eaecf-kube-api-access-wnwn4\") pod \"machine-config-controller-84d6567774-bbpqw\" (UID: \"5b9d4194-45b6-445b-b882-1982630eaecf\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bbpqw" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832664 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2390b3ae-cb4b-4e89-854e-af5c0cf04e1d-signing-cabundle\") pod \"service-ca-9c57cc56f-65h6c\" (UID: \"2390b3ae-cb4b-4e89-854e-af5c0cf04e1d\") " pod="openshift-service-ca/service-ca-9c57cc56f-65h6c" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832682 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m86j\" (UniqueName: \"kubernetes.io/projected/1df1b2d7-6ec6-4b8c-95d6-d08bdf698974-kube-api-access-4m86j\") pod \"olm-operator-6b444d44fb-9s5ck\" (UID: \"1df1b2d7-6ec6-4b8c-95d6-d08bdf698974\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9s5ck" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832699 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5b9d4194-45b6-445b-b882-1982630eaecf-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-bbpqw\" (UID: \"5b9d4194-45b6-445b-b882-1982630eaecf\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bbpqw" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832715 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91299f09-692d-4c4e-a83c-22fcd785c1ef-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832733 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmxp9\" (UniqueName: \"kubernetes.io/projected/2390b3ae-cb4b-4e89-854e-af5c0cf04e1d-kube-api-access-zmxp9\") pod \"service-ca-9c57cc56f-65h6c\" (UID: \"2390b3ae-cb4b-4e89-854e-af5c0cf04e1d\") " pod="openshift-service-ca/service-ca-9c57cc56f-65h6c" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832755 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85rms\" (UniqueName: \"kubernetes.io/projected/3239cfe6-5d28-4efc-b0b3-53699ca2a449-kube-api-access-85rms\") pod \"console-operator-58897d9998-9thjt\" (UID: \"3239cfe6-5d28-4efc-b0b3-53699ca2a449\") " pod="openshift-console-operator/console-operator-58897d9998-9thjt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832771 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1df1b2d7-6ec6-4b8c-95d6-d08bdf698974-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9s5ck\" (UID: \"1df1b2d7-6ec6-4b8c-95d6-d08bdf698974\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9s5ck" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832793 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sqnw\" (UniqueName: \"kubernetes.io/projected/aed96ecb-ad63-425d-b8f4-bd55aeb752dd-kube-api-access-4sqnw\") pod \"kube-storage-version-migrator-operator-b67b599dd-2q58n\" (UID: \"aed96ecb-ad63-425d-b8f4-bd55aeb752dd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2q58n" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832826 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/91299f09-692d-4c4e-a83c-22fcd785c1ef-encryption-config\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832844 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f62330e9-cdc5-469b-82ba-3364ce173686-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-h5tfm\" (UID: \"f62330e9-cdc5-469b-82ba-3364ce173686\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h5tfm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832865 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832881 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6fqx\" (UniqueName: \"kubernetes.io/projected/851ce288-d60d-4875-ae01-cc13268fec8c-kube-api-access-x6fqx\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.832981 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf8bw\" (UniqueName: \"kubernetes.io/projected/f62330e9-cdc5-469b-82ba-3364ce173686-kube-api-access-rf8bw\") pod \"machine-api-operator-5694c8668f-h5tfm\" (UID: \"f62330e9-cdc5-469b-82ba-3364ce173686\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h5tfm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.833026 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ad264986-2840-4a2a-bee9-9ea192cdbe86-auth-proxy-config\") pod \"machine-approver-56656f9798-khvnh\" (UID: \"ad264986-2840-4a2a-bee9-9ea192cdbe86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-khvnh" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.833057 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ed334ee3-b168-4ac1-94a7-45aa902b9b32-proxy-tls\") pod \"machine-config-operator-74547568cd-4scwl\" (UID: \"ed334ee3-b168-4ac1-94a7-45aa902b9b32\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.833123 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.833149 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4928d31-a7c1-47f0-a726-712de834e034-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-sd5f5\" (UID: \"b4928d31-a7c1-47f0-a726-712de834e034\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd5f5" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.833171 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ed334ee3-b168-4ac1-94a7-45aa902b9b32-images\") pod \"machine-config-operator-74547568cd-4scwl\" (UID: \"ed334ee3-b168-4ac1-94a7-45aa902b9b32\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.833196 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b4928d31-a7c1-47f0-a726-712de834e034-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-sd5f5\" (UID: \"b4928d31-a7c1-47f0-a726-712de834e034\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd5f5" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.833223 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f62330e9-cdc5-469b-82ba-3364ce173686-images\") pod \"machine-api-operator-5694c8668f-h5tfm\" (UID: \"f62330e9-cdc5-469b-82ba-3364ce173686\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h5tfm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.833245 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvvhj\" (UniqueName: \"kubernetes.io/projected/ad264986-2840-4a2a-bee9-9ea192cdbe86-kube-api-access-zvvhj\") pod \"machine-approver-56656f9798-khvnh\" (UID: \"ad264986-2840-4a2a-bee9-9ea192cdbe86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-khvnh" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.833282 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b83c2afe-6390-4518-a5c1-c6ca14ea4629-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5thcd\" (UID: \"b83c2afe-6390-4518-a5c1-c6ca14ea4629\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5thcd" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.834247 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b83c2afe-6390-4518-a5c1-c6ca14ea4629-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5thcd\" (UID: \"b83c2afe-6390-4518-a5c1-c6ca14ea4629\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5thcd" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.834433 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f62330e9-cdc5-469b-82ba-3364ce173686-images\") pod \"machine-api-operator-5694c8668f-h5tfm\" (UID: \"f62330e9-cdc5-469b-82ba-3364ce173686\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h5tfm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.834486 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtzxg\" (UniqueName: \"kubernetes.io/projected/b83c2afe-6390-4518-a5c1-c6ca14ea4629-kube-api-access-qtzxg\") pod \"openshift-apiserver-operator-796bbdcf4f-5thcd\" (UID: \"b83c2afe-6390-4518-a5c1-c6ca14ea4629\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5thcd" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.834526 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/44b39f4b-4213-452c-94f7-946308af1e2f-srv-cert\") pod \"catalog-operator-68c6474976-h5n9r\" (UID: \"44b39f4b-4213-452c-94f7-946308af1e2f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.834593 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/14c95e94-4284-423b-b1cf-8df8ef663d65-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qs7dj\" (UID: \"14c95e94-4284-423b-b1cf-8df8ef663d65\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qs7dj" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.834628 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/91299f09-692d-4c4e-a83c-22fcd785c1ef-audit-policies\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.834682 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b999de84-1fbb-4184-a136-ce68641fc110-config\") pod \"service-ca-operator-777779d784-f8hcm\" (UID: \"b999de84-1fbb-4184-a136-ce68641fc110\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8hcm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.834716 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/91299f09-692d-4c4e-a83c-22fcd785c1ef-audit-dir\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.834905 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ad264986-2840-4a2a-bee9-9ea192cdbe86-auth-proxy-config\") pod \"machine-approver-56656f9798-khvnh\" (UID: \"ad264986-2840-4a2a-bee9-9ea192cdbe86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-khvnh" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.835152 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad264986-2840-4a2a-bee9-9ea192cdbe86-config\") pod \"machine-approver-56656f9798-khvnh\" (UID: \"ad264986-2840-4a2a-bee9-9ea192cdbe86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-khvnh" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.835202 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aed96ecb-ad63-425d-b8f4-bd55aeb752dd-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-2q58n\" (UID: \"aed96ecb-ad63-425d-b8f4-bd55aeb752dd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2q58n" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.835239 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b55rn\" (UniqueName: \"kubernetes.io/projected/b999de84-1fbb-4184-a136-ce68641fc110-kube-api-access-b55rn\") pod \"service-ca-operator-777779d784-f8hcm\" (UID: \"b999de84-1fbb-4184-a136-ce68641fc110\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8hcm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.835295 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-audit-policies\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.835363 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx4x8\" (UniqueName: \"kubernetes.io/projected/ffb878ee-7cd7-4830-bb19-6b2a5e6bd1e0-kube-api-access-kx4x8\") pod \"multus-admission-controller-857f4d67dd-lxfpb\" (UID: \"ffb878ee-7cd7-4830-bb19-6b2a5e6bd1e0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfpb" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.835424 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f2a60c2e-1edd-4c96-b263-e84dec7ed4a2-service-ca-bundle\") pod \"authentication-operator-69f744f599-bp44p\" (UID: \"f2a60c2e-1edd-4c96-b263-e84dec7ed4a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bp44p" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.835460 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.835614 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.835849 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ad264986-2840-4a2a-bee9-9ea192cdbe86-machine-approver-tls\") pod \"machine-approver-56656f9798-khvnh\" (UID: \"ad264986-2840-4a2a-bee9-9ea192cdbe86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-khvnh" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.835886 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-audit-policies\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.836019 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.836075 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91299f09-692d-4c4e-a83c-22fcd785c1ef-serving-cert\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.836099 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f2a60c2e-1edd-4c96-b263-e84dec7ed4a2-service-ca-bundle\") pod \"authentication-operator-69f744f599-bp44p\" (UID: \"f2a60c2e-1edd-4c96-b263-e84dec7ed4a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bp44p" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.836152 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcd8k\" (UniqueName: \"kubernetes.io/projected/13b3ee79-f56b-4fff-ad26-ac72a3ef5017-kube-api-access-kcd8k\") pod \"cluster-image-registry-operator-dc59b4c8b-c9v5l\" (UID: \"13b3ee79-f56b-4fff-ad26-ac72a3ef5017\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-c9v5l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.836181 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/851ce288-d60d-4875-ae01-cc13268fec8c-audit-dir\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.836233 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/851ce288-d60d-4875-ae01-cc13268fec8c-audit-dir\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.836255 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3239cfe6-5d28-4efc-b0b3-53699ca2a449-serving-cert\") pod \"console-operator-58897d9998-9thjt\" (UID: \"3239cfe6-5d28-4efc-b0b3-53699ca2a449\") " pod="openshift-console-operator/console-operator-58897d9998-9thjt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.836297 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3239cfe6-5d28-4efc-b0b3-53699ca2a449-config\") pod \"console-operator-58897d9998-9thjt\" (UID: \"3239cfe6-5d28-4efc-b0b3-53699ca2a449\") " pod="openshift-console-operator/console-operator-58897d9998-9thjt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.836441 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55x4q\" (UniqueName: \"kubernetes.io/projected/44b39f4b-4213-452c-94f7-946308af1e2f-kube-api-access-55x4q\") pod \"catalog-operator-68c6474976-h5n9r\" (UID: \"44b39f4b-4213-452c-94f7-946308af1e2f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.836491 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.836841 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwzfz\" (UniqueName: \"kubernetes.io/projected/f2a60c2e-1edd-4c96-b263-e84dec7ed4a2-kube-api-access-gwzfz\") pod \"authentication-operator-69f744f599-bp44p\" (UID: \"f2a60c2e-1edd-4c96-b263-e84dec7ed4a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bp44p" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.836901 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3239cfe6-5d28-4efc-b0b3-53699ca2a449-config\") pod \"console-operator-58897d9998-9thjt\" (UID: \"3239cfe6-5d28-4efc-b0b3-53699ca2a449\") " pod="openshift-console-operator/console-operator-58897d9998-9thjt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.836904 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.836962 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6ef60c27-2062-4bc8-8c37-e13db97d29a3-config-volume\") pod \"collect-profiles-29568015-msx48\" (UID: \"6ef60c27-2062-4bc8-8c37-e13db97d29a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.837006 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.837052 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.837125 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/934c62b8-c7a7-4018-b5f1-7954b2baca00-apiservice-cert\") pod \"packageserver-d55dfcdfc-h7fjx\" (UID: \"934c62b8-c7a7-4018-b5f1-7954b2baca00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.837154 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzcdq\" (UniqueName: \"kubernetes.io/projected/91299f09-692d-4c4e-a83c-22fcd785c1ef-kube-api-access-rzcdq\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.837223 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f62330e9-cdc5-469b-82ba-3364ce173686-config\") pod \"machine-api-operator-5694c8668f-h5tfm\" (UID: \"f62330e9-cdc5-469b-82ba-3364ce173686\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h5tfm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.837266 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed96ecb-ad63-425d-b8f4-bd55aeb752dd-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-2q58n\" (UID: \"aed96ecb-ad63-425d-b8f4-bd55aeb752dd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2q58n" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.838008 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.838216 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/dc140eb2-a579-45bc-9fb4-247c254533ea-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qc5tq\" (UID: \"dc140eb2-a579-45bc-9fb4-247c254533ea\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qc5tq" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.838259 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6ef60c27-2062-4bc8-8c37-e13db97d29a3-secret-volume\") pod \"collect-profiles-29568015-msx48\" (UID: \"6ef60c27-2062-4bc8-8c37-e13db97d29a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.838330 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.838412 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f2a60c2e-1edd-4c96-b263-e84dec7ed4a2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bp44p\" (UID: \"f2a60c2e-1edd-4c96-b263-e84dec7ed4a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bp44p" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.838444 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f8j5\" (UniqueName: \"kubernetes.io/projected/bf190816-3a09-4918-ab57-d31e55919f3e-kube-api-access-7f8j5\") pod \"migrator-59844c95c7-59mfl\" (UID: \"bf190816-3a09-4918-ab57-d31e55919f3e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-59mfl" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.838920 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e26909f3-14e6-4f5a-9ea9-5b44ac3aa56a-metrics-tls\") pod \"dns-operator-744455d44c-6m4ch\" (UID: \"e26909f3-14e6-4f5a-9ea9-5b44ac3aa56a\") " pod="openshift-dns-operator/dns-operator-744455d44c-6m4ch" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.839002 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f2a60c2e-1edd-4c96-b263-e84dec7ed4a2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bp44p\" (UID: \"f2a60c2e-1edd-4c96-b263-e84dec7ed4a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bp44p" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.839056 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3239cfe6-5d28-4efc-b0b3-53699ca2a449-trusted-ca\") pod \"console-operator-58897d9998-9thjt\" (UID: \"3239cfe6-5d28-4efc-b0b3-53699ca2a449\") " pod="openshift-console-operator/console-operator-58897d9998-9thjt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.839085 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b83c2afe-6390-4518-a5c1-c6ca14ea4629-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5thcd\" (UID: \"b83c2afe-6390-4518-a5c1-c6ca14ea4629\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5thcd" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.839204 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.839224 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.839264 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.839494 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14c95e94-4284-423b-b1cf-8df8ef663d65-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qs7dj\" (UID: \"14c95e94-4284-423b-b1cf-8df8ef663d65\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qs7dj" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.839580 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b999de84-1fbb-4184-a136-ce68641fc110-serving-cert\") pod \"service-ca-operator-777779d784-f8hcm\" (UID: \"b999de84-1fbb-4184-a136-ce68641fc110\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8hcm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.839670 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktm4b\" (UniqueName: \"kubernetes.io/projected/ed334ee3-b168-4ac1-94a7-45aa902b9b32-kube-api-access-ktm4b\") pod \"machine-config-operator-74547568cd-4scwl\" (UID: \"ed334ee3-b168-4ac1-94a7-45aa902b9b32\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.839763 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2a60c2e-1edd-4c96-b263-e84dec7ed4a2-serving-cert\") pod \"authentication-operator-69f744f599-bp44p\" (UID: \"f2a60c2e-1edd-4c96-b263-e84dec7ed4a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bp44p" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.840070 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1df1b2d7-6ec6-4b8c-95d6-d08bdf698974-srv-cert\") pod \"olm-operator-6b444d44fb-9s5ck\" (UID: \"1df1b2d7-6ec6-4b8c-95d6-d08bdf698974\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9s5ck" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.840118 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/91299f09-692d-4c4e-a83c-22fcd785c1ef-etcd-client\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.840151 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2a60c2e-1edd-4c96-b263-e84dec7ed4a2-config\") pod \"authentication-operator-69f744f599-bp44p\" (UID: \"f2a60c2e-1edd-4c96-b263-e84dec7ed4a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bp44p" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.840183 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.840195 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.840231 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/13b3ee79-f56b-4fff-ad26-ac72a3ef5017-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-c9v5l\" (UID: \"13b3ee79-f56b-4fff-ad26-ac72a3ef5017\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-c9v5l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.840296 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j69wn\" (UniqueName: \"kubernetes.io/projected/e26909f3-14e6-4f5a-9ea9-5b44ac3aa56a-kube-api-access-j69wn\") pod \"dns-operator-744455d44c-6m4ch\" (UID: \"e26909f3-14e6-4f5a-9ea9-5b44ac3aa56a\") " pod="openshift-dns-operator/dns-operator-744455d44c-6m4ch" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.840327 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/13b3ee79-f56b-4fff-ad26-ac72a3ef5017-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-c9v5l\" (UID: \"13b3ee79-f56b-4fff-ad26-ac72a3ef5017\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-c9v5l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.840351 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.840363 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3239cfe6-5d28-4efc-b0b3-53699ca2a449-trusted-ca\") pod \"console-operator-58897d9998-9thjt\" (UID: \"3239cfe6-5d28-4efc-b0b3-53699ca2a449\") " pod="openshift-console-operator/console-operator-58897d9998-9thjt" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.840382 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2390b3ae-cb4b-4e89-854e-af5c0cf04e1d-signing-key\") pod \"service-ca-9c57cc56f-65h6c\" (UID: \"2390b3ae-cb4b-4e89-854e-af5c0cf04e1d\") " pod="openshift-service-ca/service-ca-9c57cc56f-65h6c" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.840443 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f62330e9-cdc5-469b-82ba-3364ce173686-config\") pod \"machine-api-operator-5694c8668f-h5tfm\" (UID: \"f62330e9-cdc5-469b-82ba-3364ce173686\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h5tfm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.841041 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2a60c2e-1edd-4c96-b263-e84dec7ed4a2-config\") pod \"authentication-operator-69f744f599-bp44p\" (UID: \"f2a60c2e-1edd-4c96-b263-e84dec7ed4a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bp44p" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.841224 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.841230 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/13b3ee79-f56b-4fff-ad26-ac72a3ef5017-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-c9v5l\" (UID: \"13b3ee79-f56b-4fff-ad26-ac72a3ef5017\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-c9v5l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.842051 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/dc140eb2-a579-45bc-9fb4-247c254533ea-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qc5tq\" (UID: \"dc140eb2-a579-45bc-9fb4-247c254533ea\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qc5tq" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.843262 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2a60c2e-1edd-4c96-b263-e84dec7ed4a2-serving-cert\") pod \"authentication-operator-69f744f599-bp44p\" (UID: \"f2a60c2e-1edd-4c96-b263-e84dec7ed4a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bp44p" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.843711 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b83c2afe-6390-4518-a5c1-c6ca14ea4629-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5thcd\" (UID: \"b83c2afe-6390-4518-a5c1-c6ca14ea4629\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5thcd" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.843791 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f62330e9-cdc5-469b-82ba-3364ce173686-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-h5tfm\" (UID: \"f62330e9-cdc5-469b-82ba-3364ce173686\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h5tfm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.843898 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/13b3ee79-f56b-4fff-ad26-ac72a3ef5017-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-c9v5l\" (UID: \"13b3ee79-f56b-4fff-ad26-ac72a3ef5017\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-c9v5l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.845183 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.845257 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.852419 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.856775 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.876837 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.896479 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.917011 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.937099 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.941423 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91299f09-692d-4c4e-a83c-22fcd785c1ef-serving-cert\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.941482 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55x4q\" (UniqueName: \"kubernetes.io/projected/44b39f4b-4213-452c-94f7-946308af1e2f-kube-api-access-55x4q\") pod \"catalog-operator-68c6474976-h5n9r\" (UID: \"44b39f4b-4213-452c-94f7-946308af1e2f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.941530 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6ef60c27-2062-4bc8-8c37-e13db97d29a3-config-volume\") pod \"collect-profiles-29568015-msx48\" (UID: \"6ef60c27-2062-4bc8-8c37-e13db97d29a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.941554 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/934c62b8-c7a7-4018-b5f1-7954b2baca00-apiservice-cert\") pod \"packageserver-d55dfcdfc-h7fjx\" (UID: \"934c62b8-c7a7-4018-b5f1-7954b2baca00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.941575 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzcdq\" (UniqueName: \"kubernetes.io/projected/91299f09-692d-4c4e-a83c-22fcd785c1ef-kube-api-access-rzcdq\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.941605 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed96ecb-ad63-425d-b8f4-bd55aeb752dd-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-2q58n\" (UID: \"aed96ecb-ad63-425d-b8f4-bd55aeb752dd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2q58n" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.941860 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6ef60c27-2062-4bc8-8c37-e13db97d29a3-secret-volume\") pod \"collect-profiles-29568015-msx48\" (UID: \"6ef60c27-2062-4bc8-8c37-e13db97d29a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.941909 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f8j5\" (UniqueName: \"kubernetes.io/projected/bf190816-3a09-4918-ab57-d31e55919f3e-kube-api-access-7f8j5\") pod \"migrator-59844c95c7-59mfl\" (UID: \"bf190816-3a09-4918-ab57-d31e55919f3e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-59mfl" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.941940 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14c95e94-4284-423b-b1cf-8df8ef663d65-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qs7dj\" (UID: \"14c95e94-4284-423b-b1cf-8df8ef663d65\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qs7dj" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.941976 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b999de84-1fbb-4184-a136-ce68641fc110-serving-cert\") pod \"service-ca-operator-777779d784-f8hcm\" (UID: \"b999de84-1fbb-4184-a136-ce68641fc110\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8hcm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.942008 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktm4b\" (UniqueName: \"kubernetes.io/projected/ed334ee3-b168-4ac1-94a7-45aa902b9b32-kube-api-access-ktm4b\") pod \"machine-config-operator-74547568cd-4scwl\" (UID: \"ed334ee3-b168-4ac1-94a7-45aa902b9b32\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.942037 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1df1b2d7-6ec6-4b8c-95d6-d08bdf698974-srv-cert\") pod \"olm-operator-6b444d44fb-9s5ck\" (UID: \"1df1b2d7-6ec6-4b8c-95d6-d08bdf698974\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9s5ck" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.942061 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/91299f09-692d-4c4e-a83c-22fcd785c1ef-etcd-client\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.942104 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2390b3ae-cb4b-4e89-854e-af5c0cf04e1d-signing-key\") pod \"service-ca-9c57cc56f-65h6c\" (UID: \"2390b3ae-cb4b-4e89-854e-af5c0cf04e1d\") " pod="openshift-service-ca/service-ca-9c57cc56f-65h6c" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.942140 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/934c62b8-c7a7-4018-b5f1-7954b2baca00-webhook-cert\") pod \"packageserver-d55dfcdfc-h7fjx\" (UID: \"934c62b8-c7a7-4018-b5f1-7954b2baca00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.942167 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4928d31-a7c1-47f0-a726-712de834e034-config\") pod \"kube-controller-manager-operator-78b949d7b-sd5f5\" (UID: \"b4928d31-a7c1-47f0-a726-712de834e034\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd5f5" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.942214 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh4qr\" (UniqueName: \"kubernetes.io/projected/934c62b8-c7a7-4018-b5f1-7954b2baca00-kube-api-access-mh4qr\") pod \"packageserver-d55dfcdfc-h7fjx\" (UID: \"934c62b8-c7a7-4018-b5f1-7954b2baca00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.942242 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hbtq\" (UniqueName: \"kubernetes.io/projected/6ef60c27-2062-4bc8-8c37-e13db97d29a3-kube-api-access-8hbtq\") pod \"collect-profiles-29568015-msx48\" (UID: \"6ef60c27-2062-4bc8-8c37-e13db97d29a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.942402 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5b9d4194-45b6-445b-b882-1982630eaecf-proxy-tls\") pod \"machine-config-controller-84d6567774-bbpqw\" (UID: \"5b9d4194-45b6-445b-b882-1982630eaecf\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bbpqw" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.942532 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ffb878ee-7cd7-4830-bb19-6b2a5e6bd1e0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-lxfpb\" (UID: \"ffb878ee-7cd7-4830-bb19-6b2a5e6bd1e0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfpb" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.942612 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/44b39f4b-4213-452c-94f7-946308af1e2f-profile-collector-cert\") pod \"catalog-operator-68c6474976-h5n9r\" (UID: \"44b39f4b-4213-452c-94f7-946308af1e2f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.942646 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ed334ee3-b168-4ac1-94a7-45aa902b9b32-auth-proxy-config\") pod \"machine-config-operator-74547568cd-4scwl\" (UID: \"ed334ee3-b168-4ac1-94a7-45aa902b9b32\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.942845 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/934c62b8-c7a7-4018-b5f1-7954b2baca00-tmpfs\") pod \"packageserver-d55dfcdfc-h7fjx\" (UID: \"934c62b8-c7a7-4018-b5f1-7954b2baca00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.942873 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/91299f09-692d-4c4e-a83c-22fcd785c1ef-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.942902 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4928d31-a7c1-47f0-a726-712de834e034-config\") pod \"kube-controller-manager-operator-78b949d7b-sd5f5\" (UID: \"b4928d31-a7c1-47f0-a726-712de834e034\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd5f5" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.942935 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14c95e94-4284-423b-b1cf-8df8ef663d65-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qs7dj\" (UID: \"14c95e94-4284-423b-b1cf-8df8ef663d65\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qs7dj" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.942963 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnwn4\" (UniqueName: \"kubernetes.io/projected/5b9d4194-45b6-445b-b882-1982630eaecf-kube-api-access-wnwn4\") pod \"machine-config-controller-84d6567774-bbpqw\" (UID: \"5b9d4194-45b6-445b-b882-1982630eaecf\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bbpqw" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943085 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2390b3ae-cb4b-4e89-854e-af5c0cf04e1d-signing-cabundle\") pod \"service-ca-9c57cc56f-65h6c\" (UID: \"2390b3ae-cb4b-4e89-854e-af5c0cf04e1d\") " pod="openshift-service-ca/service-ca-9c57cc56f-65h6c" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943116 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m86j\" (UniqueName: \"kubernetes.io/projected/1df1b2d7-6ec6-4b8c-95d6-d08bdf698974-kube-api-access-4m86j\") pod \"olm-operator-6b444d44fb-9s5ck\" (UID: \"1df1b2d7-6ec6-4b8c-95d6-d08bdf698974\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9s5ck" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943143 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5b9d4194-45b6-445b-b882-1982630eaecf-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-bbpqw\" (UID: \"5b9d4194-45b6-445b-b882-1982630eaecf\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bbpqw" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943163 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91299f09-692d-4c4e-a83c-22fcd785c1ef-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943186 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmxp9\" (UniqueName: \"kubernetes.io/projected/2390b3ae-cb4b-4e89-854e-af5c0cf04e1d-kube-api-access-zmxp9\") pod \"service-ca-9c57cc56f-65h6c\" (UID: \"2390b3ae-cb4b-4e89-854e-af5c0cf04e1d\") " pod="openshift-service-ca/service-ca-9c57cc56f-65h6c" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943295 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1df1b2d7-6ec6-4b8c-95d6-d08bdf698974-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9s5ck\" (UID: \"1df1b2d7-6ec6-4b8c-95d6-d08bdf698974\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9s5ck" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943349 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sqnw\" (UniqueName: \"kubernetes.io/projected/aed96ecb-ad63-425d-b8f4-bd55aeb752dd-kube-api-access-4sqnw\") pod \"kube-storage-version-migrator-operator-b67b599dd-2q58n\" (UID: \"aed96ecb-ad63-425d-b8f4-bd55aeb752dd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2q58n" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943380 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/91299f09-692d-4c4e-a83c-22fcd785c1ef-encryption-config\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943441 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ed334ee3-b168-4ac1-94a7-45aa902b9b32-proxy-tls\") pod \"machine-config-operator-74547568cd-4scwl\" (UID: \"ed334ee3-b168-4ac1-94a7-45aa902b9b32\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943469 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4928d31-a7c1-47f0-a726-712de834e034-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-sd5f5\" (UID: \"b4928d31-a7c1-47f0-a726-712de834e034\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd5f5" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943491 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ed334ee3-b168-4ac1-94a7-45aa902b9b32-images\") pod \"machine-config-operator-74547568cd-4scwl\" (UID: \"ed334ee3-b168-4ac1-94a7-45aa902b9b32\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943517 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b4928d31-a7c1-47f0-a726-712de834e034-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-sd5f5\" (UID: \"b4928d31-a7c1-47f0-a726-712de834e034\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd5f5" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943570 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/44b39f4b-4213-452c-94f7-946308af1e2f-srv-cert\") pod \"catalog-operator-68c6474976-h5n9r\" (UID: \"44b39f4b-4213-452c-94f7-946308af1e2f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943604 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/14c95e94-4284-423b-b1cf-8df8ef663d65-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qs7dj\" (UID: \"14c95e94-4284-423b-b1cf-8df8ef663d65\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qs7dj" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943609 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/934c62b8-c7a7-4018-b5f1-7954b2baca00-tmpfs\") pod \"packageserver-d55dfcdfc-h7fjx\" (UID: \"934c62b8-c7a7-4018-b5f1-7954b2baca00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943630 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/91299f09-692d-4c4e-a83c-22fcd785c1ef-audit-policies\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943653 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b999de84-1fbb-4184-a136-ce68641fc110-config\") pod \"service-ca-operator-777779d784-f8hcm\" (UID: \"b999de84-1fbb-4184-a136-ce68641fc110\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8hcm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943673 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/91299f09-692d-4c4e-a83c-22fcd785c1ef-audit-dir\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943695 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aed96ecb-ad63-425d-b8f4-bd55aeb752dd-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-2q58n\" (UID: \"aed96ecb-ad63-425d-b8f4-bd55aeb752dd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2q58n" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943721 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b55rn\" (UniqueName: \"kubernetes.io/projected/b999de84-1fbb-4184-a136-ce68641fc110-kube-api-access-b55rn\") pod \"service-ca-operator-777779d784-f8hcm\" (UID: \"b999de84-1fbb-4184-a136-ce68641fc110\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8hcm" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943749 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx4x8\" (UniqueName: \"kubernetes.io/projected/ffb878ee-7cd7-4830-bb19-6b2a5e6bd1e0-kube-api-access-kx4x8\") pod \"multus-admission-controller-857f4d67dd-lxfpb\" (UID: \"ffb878ee-7cd7-4830-bb19-6b2a5e6bd1e0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfpb" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943871 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ed334ee3-b168-4ac1-94a7-45aa902b9b32-auth-proxy-config\") pod \"machine-config-operator-74547568cd-4scwl\" (UID: \"ed334ee3-b168-4ac1-94a7-45aa902b9b32\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.943894 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/91299f09-692d-4c4e-a83c-22fcd785c1ef-audit-dir\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.944387 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/91299f09-692d-4c4e-a83c-22fcd785c1ef-audit-policies\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.944457 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5b9d4194-45b6-445b-b882-1982630eaecf-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-bbpqw\" (UID: \"5b9d4194-45b6-445b-b882-1982630eaecf\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bbpqw" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.945337 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/91299f09-692d-4c4e-a83c-22fcd785c1ef-etcd-client\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.946773 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4928d31-a7c1-47f0-a726-712de834e034-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-sd5f5\" (UID: \"b4928d31-a7c1-47f0-a726-712de834e034\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd5f5" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.957163 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.964299 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91299f09-692d-4c4e-a83c-22fcd785c1ef-serving-cert\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.976399 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.985905 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/91299f09-692d-4c4e-a83c-22fcd785c1ef-encryption-config\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:41 crc kubenswrapper[4696]: I0321 08:29:41.997454 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.004234 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/91299f09-692d-4c4e-a83c-22fcd785c1ef-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.017635 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.023914 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91299f09-692d-4c4e-a83c-22fcd785c1ef-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.036270 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.060371 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.076908 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.088013 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aed96ecb-ad63-425d-b8f4-bd55aeb752dd-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-2q58n\" (UID: \"aed96ecb-ad63-425d-b8f4-bd55aeb752dd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2q58n" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.096956 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.117126 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.137159 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.142738 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed96ecb-ad63-425d-b8f4-bd55aeb752dd-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-2q58n\" (UID: \"aed96ecb-ad63-425d-b8f4-bd55aeb752dd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2q58n" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.157203 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.176521 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.188115 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ffb878ee-7cd7-4830-bb19-6b2a5e6bd1e0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-lxfpb\" (UID: \"ffb878ee-7cd7-4830-bb19-6b2a5e6bd1e0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfpb" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.196846 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.216638 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.238398 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.257106 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.277025 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.297094 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.316755 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.337844 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.356344 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.377156 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.386310 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14c95e94-4284-423b-b1cf-8df8ef663d65-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qs7dj\" (UID: \"14c95e94-4284-423b-b1cf-8df8ef663d65\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qs7dj" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.397313 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.403000 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14c95e94-4284-423b-b1cf-8df8ef663d65-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qs7dj\" (UID: \"14c95e94-4284-423b-b1cf-8df8ef663d65\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qs7dj" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.417187 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.437507 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.457876 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.477542 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.497274 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.518187 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.534157 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w95sn" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.536762 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.557533 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.576079 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.597264 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.617725 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.637121 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.662420 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.677385 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.695991 4696 request.go:700] Waited for 1.002593571s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.697680 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.717167 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.737036 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.764896 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.777361 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.797071 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.817796 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.837139 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.857726 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.877357 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.885673 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5b9d4194-45b6-445b-b882-1982630eaecf-proxy-tls\") pod \"machine-config-controller-84d6567774-bbpqw\" (UID: \"5b9d4194-45b6-445b-b882-1982630eaecf\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bbpqw" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.896677 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.916626 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.925251 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ed334ee3-b168-4ac1-94a7-45aa902b9b32-images\") pod \"machine-config-operator-74547568cd-4scwl\" (UID: \"ed334ee3-b168-4ac1-94a7-45aa902b9b32\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl" Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.937584 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.941763 4696 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.941910 4696 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.941935 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/934c62b8-c7a7-4018-b5f1-7954b2baca00-apiservice-cert podName:934c62b8-c7a7-4018-b5f1-7954b2baca00 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:43.44190877 +0000 UTC m=+117.562789483 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/934c62b8-c7a7-4018-b5f1-7954b2baca00-apiservice-cert") pod "packageserver-d55dfcdfc-h7fjx" (UID: "934c62b8-c7a7-4018-b5f1-7954b2baca00") : failed to sync secret cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.942087 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6ef60c27-2062-4bc8-8c37-e13db97d29a3-config-volume podName:6ef60c27-2062-4bc8-8c37-e13db97d29a3 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:43.442057433 +0000 UTC m=+117.562938176 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/6ef60c27-2062-4bc8-8c37-e13db97d29a3-config-volume") pod "collect-profiles-29568015-msx48" (UID: "6ef60c27-2062-4bc8-8c37-e13db97d29a3") : failed to sync configmap cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.942131 4696 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.942167 4696 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.942215 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b999de84-1fbb-4184-a136-ce68641fc110-serving-cert podName:b999de84-1fbb-4184-a136-ce68641fc110 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:43.442189608 +0000 UTC m=+117.563070361 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/b999de84-1fbb-4184-a136-ce68641fc110-serving-cert") pod "service-ca-operator-777779d784-f8hcm" (UID: "b999de84-1fbb-4184-a136-ce68641fc110") : failed to sync secret cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.942131 4696 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.942253 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1df1b2d7-6ec6-4b8c-95d6-d08bdf698974-srv-cert podName:1df1b2d7-6ec6-4b8c-95d6-d08bdf698974 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:43.442233029 +0000 UTC m=+117.563113772 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/1df1b2d7-6ec6-4b8c-95d6-d08bdf698974-srv-cert") pod "olm-operator-6b444d44fb-9s5ck" (UID: "1df1b2d7-6ec6-4b8c-95d6-d08bdf698974") : failed to sync secret cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.942275 4696 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.942277 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6ef60c27-2062-4bc8-8c37-e13db97d29a3-secret-volume podName:6ef60c27-2062-4bc8-8c37-e13db97d29a3 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:43.44226627 +0000 UTC m=+117.563147023 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-volume" (UniqueName: "kubernetes.io/secret/6ef60c27-2062-4bc8-8c37-e13db97d29a3-secret-volume") pod "collect-profiles-29568015-msx48" (UID: "6ef60c27-2062-4bc8-8c37-e13db97d29a3") : failed to sync secret cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.942282 4696 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.942298 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/934c62b8-c7a7-4018-b5f1-7954b2baca00-webhook-cert podName:934c62b8-c7a7-4018-b5f1-7954b2baca00 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:43.44229007 +0000 UTC m=+117.563170783 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/934c62b8-c7a7-4018-b5f1-7954b2baca00-webhook-cert") pod "packageserver-d55dfcdfc-h7fjx" (UID: "934c62b8-c7a7-4018-b5f1-7954b2baca00") : failed to sync secret cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.942350 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2390b3ae-cb4b-4e89-854e-af5c0cf04e1d-signing-key podName:2390b3ae-cb4b-4e89-854e-af5c0cf04e1d nodeName:}" failed. No retries permitted until 2026-03-21 08:29:43.442332461 +0000 UTC m=+117.563213264 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/2390b3ae-cb4b-4e89-854e-af5c0cf04e1d-signing-key") pod "service-ca-9c57cc56f-65h6c" (UID: "2390b3ae-cb4b-4e89-854e-af5c0cf04e1d") : failed to sync secret cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.943402 4696 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.943438 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/44b39f4b-4213-452c-94f7-946308af1e2f-profile-collector-cert podName:44b39f4b-4213-452c-94f7-946308af1e2f nodeName:}" failed. No retries permitted until 2026-03-21 08:29:43.443429189 +0000 UTC m=+117.564309892 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/44b39f4b-4213-452c-94f7-946308af1e2f-profile-collector-cert") pod "catalog-operator-68c6474976-h5n9r" (UID: "44b39f4b-4213-452c-94f7-946308af1e2f") : failed to sync secret cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.943462 4696 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.943483 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1df1b2d7-6ec6-4b8c-95d6-d08bdf698974-profile-collector-cert podName:1df1b2d7-6ec6-4b8c-95d6-d08bdf698974 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:43.4434778 +0000 UTC m=+117.564358503 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/1df1b2d7-6ec6-4b8c-95d6-d08bdf698974-profile-collector-cert") pod "olm-operator-6b444d44fb-9s5ck" (UID: "1df1b2d7-6ec6-4b8c-95d6-d08bdf698974") : failed to sync secret cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.943532 4696 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.943725 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2390b3ae-cb4b-4e89-854e-af5c0cf04e1d-signing-cabundle podName:2390b3ae-cb4b-4e89-854e-af5c0cf04e1d nodeName:}" failed. No retries permitted until 2026-03-21 08:29:43.443685636 +0000 UTC m=+117.564566399 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/2390b3ae-cb4b-4e89-854e-af5c0cf04e1d-signing-cabundle") pod "service-ca-9c57cc56f-65h6c" (UID: "2390b3ae-cb4b-4e89-854e-af5c0cf04e1d") : failed to sync configmap cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.943867 4696 secret.go:188] Couldn't get secret openshift-machine-config-operator/mco-proxy-tls: failed to sync secret cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.943968 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ed334ee3-b168-4ac1-94a7-45aa902b9b32-proxy-tls podName:ed334ee3-b168-4ac1-94a7-45aa902b9b32 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:43.443928073 +0000 UTC m=+117.564808846 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/ed334ee3-b168-4ac1-94a7-45aa902b9b32-proxy-tls") pod "machine-config-operator-74547568cd-4scwl" (UID: "ed334ee3-b168-4ac1-94a7-45aa902b9b32") : failed to sync secret cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.943966 4696 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.943995 4696 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.944033 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b999de84-1fbb-4184-a136-ce68641fc110-config podName:b999de84-1fbb-4184-a136-ce68641fc110 nodeName:}" failed. No retries permitted until 2026-03-21 08:29:43.444016675 +0000 UTC m=+117.564897478 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/b999de84-1fbb-4184-a136-ce68641fc110-config") pod "service-ca-operator-777779d784-f8hcm" (UID: "b999de84-1fbb-4184-a136-ce68641fc110") : failed to sync configmap cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: E0321 08:29:42.944071 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/44b39f4b-4213-452c-94f7-946308af1e2f-srv-cert podName:44b39f4b-4213-452c-94f7-946308af1e2f nodeName:}" failed. No retries permitted until 2026-03-21 08:29:43.444047676 +0000 UTC m=+117.564928479 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/44b39f4b-4213-452c-94f7-946308af1e2f-srv-cert") pod "catalog-operator-68c6474976-h5n9r" (UID: "44b39f4b-4213-452c-94f7-946308af1e2f") : failed to sync secret cache: timed out waiting for the condition Mar 21 08:29:42 crc kubenswrapper[4696]: I0321 08:29:42.957210 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.013535 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvhrn\" (UniqueName: \"kubernetes.io/projected/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-kube-api-access-qvhrn\") pod \"controller-manager-879f6c89f-ptg5t\" (UID: \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.034552 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nspjm\" (UniqueName: \"kubernetes.io/projected/68a1b8db-aec4-4681-b1f0-87761e34c518-kube-api-access-nspjm\") pod \"openshift-config-operator-7777fb866f-xt5jj\" (UID: \"68a1b8db-aec4-4681-b1f0-87761e34c518\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xt5jj" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.043646 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.051129 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48rwn\" (UniqueName: \"kubernetes.io/projected/78c4d6cc-8ad5-4d4b-b20d-bebe8259e18e-kube-api-access-48rwn\") pod \"openshift-controller-manager-operator-756b6f6bc6-vnbgm\" (UID: \"78c4d6cc-8ad5-4d4b-b20d-bebe8259e18e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vnbgm" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.071642 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr74r\" (UniqueName: \"kubernetes.io/projected/fb015b47-dbed-4cc5-a8d4-ab5d55eb1514-kube-api-access-lr74r\") pod \"etcd-operator-b45778765-rvwvm\" (UID: \"fb015b47-dbed-4cc5-a8d4-ab5d55eb1514\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.082075 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.091805 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9q56\" (UniqueName: \"kubernetes.io/projected/d1d05966-d105-4c45-bfa2-e66a67882baf-kube-api-access-x9q56\") pod \"console-f9d7485db-f5n2g\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.108257 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xt5jj" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.112325 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6v76\" (UniqueName: \"kubernetes.io/projected/e24de32f-7778-4fac-815b-fa4d2a44dd2f-kube-api-access-v6v76\") pod \"apiserver-76f77b778f-xblff\" (UID: \"e24de32f-7778-4fac-815b-fa4d2a44dd2f\") " pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.136101 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc6gp\" (UniqueName: \"kubernetes.io/projected/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-kube-api-access-lc6gp\") pod \"route-controller-manager-6576b87f9c-xnns6\" (UID: \"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.137960 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.139530 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.157336 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.163726 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.177294 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.184216 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vnbgm" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.198836 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.217978 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.236916 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.256180 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ptg5t"] Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.257436 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 21 08:29:43 crc kubenswrapper[4696]: W0321 08:29:43.273952 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14fd07c2_17e1_4adb_87c8_459f1dcb41c2.slice/crio-1b330f2eb9e11f0e261c92ff501074e176cbf65d9bb598aa233d095589c70fe1 WatchSource:0}: Error finding container 1b330f2eb9e11f0e261c92ff501074e176cbf65d9bb598aa233d095589c70fe1: Status 404 returned error can't find the container with id 1b330f2eb9e11f0e261c92ff501074e176cbf65d9bb598aa233d095589c70fe1 Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.277574 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.296884 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.317843 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.338565 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.358021 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.365926 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.368666 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:43 crc kubenswrapper[4696]: E0321 08:29:43.369365 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:30:15.36934483 +0000 UTC m=+149.490225543 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.380702 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.385131 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-xblff"] Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.395605 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6"] Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.397077 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 21 08:29:43 crc kubenswrapper[4696]: W0321 08:29:43.397592 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode24de32f_7778_4fac_815b_fa4d2a44dd2f.slice/crio-5634def7b17842ce1f4a3c2a132bacd3abb6fa259a67b7afd1e0a0ba58eed81f WatchSource:0}: Error finding container 5634def7b17842ce1f4a3c2a132bacd3abb6fa259a67b7afd1e0a0ba58eed81f: Status 404 returned error can't find the container with id 5634def7b17842ce1f4a3c2a132bacd3abb6fa259a67b7afd1e0a0ba58eed81f Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.416321 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.424081 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vnbgm"] Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.446132 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.456419 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.469922 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1df1b2d7-6ec6-4b8c-95d6-d08bdf698974-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9s5ck\" (UID: \"1df1b2d7-6ec6-4b8c-95d6-d08bdf698974\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9s5ck" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.469981 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.470005 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ed334ee3-b168-4ac1-94a7-45aa902b9b32-proxy-tls\") pod \"machine-config-operator-74547568cd-4scwl\" (UID: \"ed334ee3-b168-4ac1-94a7-45aa902b9b32\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.470045 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/44b39f4b-4213-452c-94f7-946308af1e2f-srv-cert\") pod \"catalog-operator-68c6474976-h5n9r\" (UID: \"44b39f4b-4213-452c-94f7-946308af1e2f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.470827 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b999de84-1fbb-4184-a136-ce68641fc110-config\") pod \"service-ca-operator-777779d784-f8hcm\" (UID: \"b999de84-1fbb-4184-a136-ce68641fc110\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8hcm" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.470875 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.470927 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6ef60c27-2062-4bc8-8c37-e13db97d29a3-config-volume\") pod \"collect-profiles-29568015-msx48\" (UID: \"6ef60c27-2062-4bc8-8c37-e13db97d29a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.470947 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/934c62b8-c7a7-4018-b5f1-7954b2baca00-apiservice-cert\") pod \"packageserver-d55dfcdfc-h7fjx\" (UID: \"934c62b8-c7a7-4018-b5f1-7954b2baca00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.470975 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6ef60c27-2062-4bc8-8c37-e13db97d29a3-secret-volume\") pod \"collect-profiles-29568015-msx48\" (UID: \"6ef60c27-2062-4bc8-8c37-e13db97d29a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.471004 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b999de84-1fbb-4184-a136-ce68641fc110-serving-cert\") pod \"service-ca-operator-777779d784-f8hcm\" (UID: \"b999de84-1fbb-4184-a136-ce68641fc110\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8hcm" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.471028 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1df1b2d7-6ec6-4b8c-95d6-d08bdf698974-srv-cert\") pod \"olm-operator-6b444d44fb-9s5ck\" (UID: \"1df1b2d7-6ec6-4b8c-95d6-d08bdf698974\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9s5ck" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.471140 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2390b3ae-cb4b-4e89-854e-af5c0cf04e1d-signing-key\") pod \"service-ca-9c57cc56f-65h6c\" (UID: \"2390b3ae-cb4b-4e89-854e-af5c0cf04e1d\") " pod="openshift-service-ca/service-ca-9c57cc56f-65h6c" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.471204 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.471260 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/934c62b8-c7a7-4018-b5f1-7954b2baca00-webhook-cert\") pod \"packageserver-d55dfcdfc-h7fjx\" (UID: \"934c62b8-c7a7-4018-b5f1-7954b2baca00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.471294 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.471315 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/44b39f4b-4213-452c-94f7-946308af1e2f-profile-collector-cert\") pod \"catalog-operator-68c6474976-h5n9r\" (UID: \"44b39f4b-4213-452c-94f7-946308af1e2f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.471351 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2390b3ae-cb4b-4e89-854e-af5c0cf04e1d-signing-cabundle\") pod \"service-ca-9c57cc56f-65h6c\" (UID: \"2390b3ae-cb4b-4e89-854e-af5c0cf04e1d\") " pod="openshift-service-ca/service-ca-9c57cc56f-65h6c" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.472062 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2390b3ae-cb4b-4e89-854e-af5c0cf04e1d-signing-cabundle\") pod \"service-ca-9c57cc56f-65h6c\" (UID: \"2390b3ae-cb4b-4e89-854e-af5c0cf04e1d\") " pod="openshift-service-ca/service-ca-9c57cc56f-65h6c" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.472511 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b999de84-1fbb-4184-a136-ce68641fc110-config\") pod \"service-ca-operator-777779d784-f8hcm\" (UID: \"b999de84-1fbb-4184-a136-ce68641fc110\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8hcm" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.474206 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6ef60c27-2062-4bc8-8c37-e13db97d29a3-config-volume\") pod \"collect-profiles-29568015-msx48\" (UID: \"6ef60c27-2062-4bc8-8c37-e13db97d29a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.476932 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ed334ee3-b168-4ac1-94a7-45aa902b9b32-proxy-tls\") pod \"machine-config-operator-74547568cd-4scwl\" (UID: \"ed334ee3-b168-4ac1-94a7-45aa902b9b32\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.482188 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1df1b2d7-6ec6-4b8c-95d6-d08bdf698974-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9s5ck\" (UID: \"1df1b2d7-6ec6-4b8c-95d6-d08bdf698974\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9s5ck" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.482982 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/934c62b8-c7a7-4018-b5f1-7954b2baca00-webhook-cert\") pod \"packageserver-d55dfcdfc-h7fjx\" (UID: \"934c62b8-c7a7-4018-b5f1-7954b2baca00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.482994 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.483986 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/934c62b8-c7a7-4018-b5f1-7954b2baca00-apiservice-cert\") pod \"packageserver-d55dfcdfc-h7fjx\" (UID: \"934c62b8-c7a7-4018-b5f1-7954b2baca00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.485223 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.485763 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2390b3ae-cb4b-4e89-854e-af5c0cf04e1d-signing-key\") pod \"service-ca-9c57cc56f-65h6c\" (UID: \"2390b3ae-cb4b-4e89-854e-af5c0cf04e1d\") " pod="openshift-service-ca/service-ca-9c57cc56f-65h6c" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.485879 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.486351 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b999de84-1fbb-4184-a136-ce68641fc110-serving-cert\") pod \"service-ca-operator-777779d784-f8hcm\" (UID: \"b999de84-1fbb-4184-a136-ce68641fc110\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8hcm" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.486438 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.486907 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1df1b2d7-6ec6-4b8c-95d6-d08bdf698974-srv-cert\") pod \"olm-operator-6b444d44fb-9s5ck\" (UID: \"1df1b2d7-6ec6-4b8c-95d6-d08bdf698974\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9s5ck" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.487329 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.489217 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/44b39f4b-4213-452c-94f7-946308af1e2f-profile-collector-cert\") pod \"catalog-operator-68c6474976-h5n9r\" (UID: \"44b39f4b-4213-452c-94f7-946308af1e2f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.490000 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6ef60c27-2062-4bc8-8c37-e13db97d29a3-secret-volume\") pod \"collect-profiles-29568015-msx48\" (UID: \"6ef60c27-2062-4bc8-8c37-e13db97d29a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.497495 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.499040 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.509793 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.517737 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.537195 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.537369 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-xt5jj"] Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.537939 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rvwvm"] Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.556908 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 21 08:29:43 crc kubenswrapper[4696]: W0321 08:29:43.560419 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb015b47_dbed_4cc5_a8d4_ab5d55eb1514.slice/crio-2830003896d36638549bc14e813db85677236b819e0f5b4936c63064dc1dfab3 WatchSource:0}: Error finding container 2830003896d36638549bc14e813db85677236b819e0f5b4936c63064dc1dfab3: Status 404 returned error can't find the container with id 2830003896d36638549bc14e813db85677236b819e0f5b4936c63064dc1dfab3 Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.565560 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 21 08:29:43 crc kubenswrapper[4696]: W0321 08:29:43.567493 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68a1b8db_aec4_4681_b1f0_87761e34c518.slice/crio-6d15f4d19e90b31127a5f9f31e4b21294721aa65117786402fbc6b3fd6cb438d WatchSource:0}: Error finding container 6d15f4d19e90b31127a5f9f31e4b21294721aa65117786402fbc6b3fd6cb438d: Status 404 returned error can't find the container with id 6d15f4d19e90b31127a5f9f31e4b21294721aa65117786402fbc6b3fd6cb438d Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.567778 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/44b39f4b-4213-452c-94f7-946308af1e2f-srv-cert\") pod \"catalog-operator-68c6474976-h5n9r\" (UID: \"44b39f4b-4213-452c-94f7-946308af1e2f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.577595 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.583292 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-f5n2g"] Mar 21 08:29:43 crc kubenswrapper[4696]: W0321 08:29:43.596793 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1d05966_d105_4c45_bfa2_e66a67882baf.slice/crio-e8ba5226619f2620d146678c29fc7042c70e2dbf434d91f6483bbbba2262dc44 WatchSource:0}: Error finding container e8ba5226619f2620d146678c29fc7042c70e2dbf434d91f6483bbbba2262dc44: Status 404 returned error can't find the container with id e8ba5226619f2620d146678c29fc7042c70e2dbf434d91f6483bbbba2262dc44 Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.623353 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.638773 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.658109 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.682093 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.697661 4696 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.716009 4696 request.go:700] Waited for 1.922688117s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.717375 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.737911 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-sysctl-allowlist" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.758442 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.782260 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.797306 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.818567 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.851401 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r74wp\" (UniqueName: \"kubernetes.io/projected/dc140eb2-a579-45bc-9fb4-247c254533ea-kube-api-access-r74wp\") pod \"cluster-samples-operator-665b6dd947-qc5tq\" (UID: \"dc140eb2-a579-45bc-9fb4-247c254533ea\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qc5tq" Mar 21 08:29:43 crc kubenswrapper[4696]: W0321 08:29:43.866328 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-295af980d87a5050fd6fca7f648de5420a8050fe50af6c9cda169972d10f19bf WatchSource:0}: Error finding container 295af980d87a5050fd6fca7f648de5420a8050fe50af6c9cda169972d10f19bf: Status 404 returned error can't find the container with id 295af980d87a5050fd6fca7f648de5420a8050fe50af6c9cda169972d10f19bf Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.871302 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/13b3ee79-f56b-4fff-ad26-ac72a3ef5017-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-c9v5l\" (UID: \"13b3ee79-f56b-4fff-ad26-ac72a3ef5017\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-c9v5l" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.890952 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85rms\" (UniqueName: \"kubernetes.io/projected/3239cfe6-5d28-4efc-b0b3-53699ca2a449-kube-api-access-85rms\") pod \"console-operator-58897d9998-9thjt\" (UID: \"3239cfe6-5d28-4efc-b0b3-53699ca2a449\") " pod="openshift-console-operator/console-operator-58897d9998-9thjt" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.893664 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-9thjt" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.913104 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6fqx\" (UniqueName: \"kubernetes.io/projected/851ce288-d60d-4875-ae01-cc13268fec8c-kube-api-access-x6fqx\") pod \"oauth-openshift-558db77b4-qvm82\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.932946 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf8bw\" (UniqueName: \"kubernetes.io/projected/f62330e9-cdc5-469b-82ba-3364ce173686-kube-api-access-rf8bw\") pod \"machine-api-operator-5694c8668f-h5tfm\" (UID: \"f62330e9-cdc5-469b-82ba-3364ce173686\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h5tfm" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.949954 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtzxg\" (UniqueName: \"kubernetes.io/projected/b83c2afe-6390-4518-a5c1-c6ca14ea4629-kube-api-access-qtzxg\") pod \"openshift-apiserver-operator-796bbdcf4f-5thcd\" (UID: \"b83c2afe-6390-4518-a5c1-c6ca14ea4629\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5thcd" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.966576 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qc5tq" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.978037 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvvhj\" (UniqueName: \"kubernetes.io/projected/ad264986-2840-4a2a-bee9-9ea192cdbe86-kube-api-access-zvvhj\") pod \"machine-approver-56656f9798-khvnh\" (UID: \"ad264986-2840-4a2a-bee9-9ea192cdbe86\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-khvnh" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.985233 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.990950 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-h5tfm" Mar 21 08:29:43 crc kubenswrapper[4696]: I0321 08:29:43.996137 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcd8k\" (UniqueName: \"kubernetes.io/projected/13b3ee79-f56b-4fff-ad26-ac72a3ef5017-kube-api-access-kcd8k\") pod \"cluster-image-registry-operator-dc59b4c8b-c9v5l\" (UID: \"13b3ee79-f56b-4fff-ad26-ac72a3ef5017\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-c9v5l" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.014896 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwzfz\" (UniqueName: \"kubernetes.io/projected/f2a60c2e-1edd-4c96-b263-e84dec7ed4a2-kube-api-access-gwzfz\") pod \"authentication-operator-69f744f599-bp44p\" (UID: \"f2a60c2e-1edd-4c96-b263-e84dec7ed4a2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bp44p" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.036690 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j69wn\" (UniqueName: \"kubernetes.io/projected/e26909f3-14e6-4f5a-9ea9-5b44ac3aa56a-kube-api-access-j69wn\") pod \"dns-operator-744455d44c-6m4ch\" (UID: \"e26909f3-14e6-4f5a-9ea9-5b44ac3aa56a\") " pod="openshift-dns-operator/dns-operator-744455d44c-6m4ch" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.056596 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55x4q\" (UniqueName: \"kubernetes.io/projected/44b39f4b-4213-452c-94f7-946308af1e2f-kube-api-access-55x4q\") pod \"catalog-operator-68c6474976-h5n9r\" (UID: \"44b39f4b-4213-452c-94f7-946308af1e2f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.077441 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzcdq\" (UniqueName: \"kubernetes.io/projected/91299f09-692d-4c4e-a83c-22fcd785c1ef-kube-api-access-rzcdq\") pod \"apiserver-7bbb656c7d-gdp2l\" (UID: \"91299f09-692d-4c4e-a83c-22fcd785c1ef\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.095318 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-9thjt"] Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.096494 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f8j5\" (UniqueName: \"kubernetes.io/projected/bf190816-3a09-4918-ab57-d31e55919f3e-kube-api-access-7f8j5\") pod \"migrator-59844c95c7-59mfl\" (UID: \"bf190816-3a09-4918-ab57-d31e55919f3e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-59mfl" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.119615 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktm4b\" (UniqueName: \"kubernetes.io/projected/ed334ee3-b168-4ac1-94a7-45aa902b9b32-kube-api-access-ktm4b\") pod \"machine-config-operator-74547568cd-4scwl\" (UID: \"ed334ee3-b168-4ac1-94a7-45aa902b9b32\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.123355 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vnbgm" event={"ID":"78c4d6cc-8ad5-4d4b-b20d-bebe8259e18e","Type":"ContainerStarted","Data":"7b6d862ec8b54850030b2ae5af9ca81a147a7369e87f1bf78805b022a658f42c"} Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.123415 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vnbgm" event={"ID":"78c4d6cc-8ad5-4d4b-b20d-bebe8259e18e","Type":"ContainerStarted","Data":"a8f5a991908439757aba5ba5d502607daa34cb00ec9d18f4a24b4b3972929f1c"} Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.127230 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" event={"ID":"fb015b47-dbed-4cc5-a8d4-ab5d55eb1514","Type":"ContainerStarted","Data":"88481dc5ef7be46e1dd8c8e139b5022ba92c3526922f676e83a74f3d8013adc3"} Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.127264 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" event={"ID":"fb015b47-dbed-4cc5-a8d4-ab5d55eb1514","Type":"ContainerStarted","Data":"2830003896d36638549bc14e813db85677236b819e0f5b4936c63064dc1dfab3"} Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.134005 4696 generic.go:334] "Generic (PLEG): container finished" podID="e24de32f-7778-4fac-815b-fa4d2a44dd2f" containerID="887b42c7048f89a6ddc6f3c1c0991d7fd543a8020b3612560c0579a0c8db9c84" exitCode=0 Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.134657 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-xblff" event={"ID":"e24de32f-7778-4fac-815b-fa4d2a44dd2f","Type":"ContainerDied","Data":"887b42c7048f89a6ddc6f3c1c0991d7fd543a8020b3612560c0579a0c8db9c84"} Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.134722 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-xblff" event={"ID":"e24de32f-7778-4fac-815b-fa4d2a44dd2f","Type":"ContainerStarted","Data":"5634def7b17842ce1f4a3c2a132bacd3abb6fa259a67b7afd1e0a0ba58eed81f"} Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.137874 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh4qr\" (UniqueName: \"kubernetes.io/projected/934c62b8-c7a7-4018-b5f1-7954b2baca00-kube-api-access-mh4qr\") pod \"packageserver-d55dfcdfc-h7fjx\" (UID: \"934c62b8-c7a7-4018-b5f1-7954b2baca00\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.141642 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" event={"ID":"14fd07c2-17e1-4adb-87c8-459f1dcb41c2","Type":"ContainerStarted","Data":"65eadb3b7870c00e93eedf571b1bfa6163cf9a08c6fadf8c64efba6e41e33426"} Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.142070 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" event={"ID":"14fd07c2-17e1-4adb-87c8-459f1dcb41c2","Type":"ContainerStarted","Data":"1b330f2eb9e11f0e261c92ff501074e176cbf65d9bb598aa233d095589c70fe1"} Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.142230 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.144255 4696 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-ptg5t container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.144293 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" podUID="14fd07c2-17e1-4adb-87c8-459f1dcb41c2" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.145643 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"b9282b9c3bb98a47c4b4dfdc282362768484d7e28dad6c34f775b1548fc91029"} Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.145693 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"295af980d87a5050fd6fca7f648de5420a8050fe50af6c9cda169972d10f19bf"} Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.155173 4696 generic.go:334] "Generic (PLEG): container finished" podID="68a1b8db-aec4-4681-b1f0-87761e34c518" containerID="16a54e0a5f9b6e4311fd068f7b0d05b8080d6eebc142efd972fc37d0578e29bf" exitCode=0 Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.155254 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xt5jj" event={"ID":"68a1b8db-aec4-4681-b1f0-87761e34c518","Type":"ContainerDied","Data":"16a54e0a5f9b6e4311fd068f7b0d05b8080d6eebc142efd972fc37d0578e29bf"} Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.155286 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xt5jj" event={"ID":"68a1b8db-aec4-4681-b1f0-87761e34c518","Type":"ContainerStarted","Data":"6d15f4d19e90b31127a5f9f31e4b21294721aa65117786402fbc6b3fd6cb438d"} Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.158430 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ee9b5463783c42ca8b9a5aa1ff5a7da6ebc6499cb74146900c6fadc31f341389"} Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.158468 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"8aeeb88aef61f505c1ca45727558089327adf73b7a2b2f37217dcf440f6b931c"} Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.160660 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hbtq\" (UniqueName: \"kubernetes.io/projected/6ef60c27-2062-4bc8-8c37-e13db97d29a3-kube-api-access-8hbtq\") pod \"collect-profiles-29568015-msx48\" (UID: \"6ef60c27-2062-4bc8-8c37-e13db97d29a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.168556 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" event={"ID":"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c","Type":"ContainerStarted","Data":"3e773b872d98df4836577cc15d9015665074913bd0488072a4a9e16c3145fe27"} Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.168591 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" event={"ID":"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c","Type":"ContainerStarted","Data":"5846194320be0a87de6399288a8b8b403658935759d629e5c55cbad3ebadc9de"} Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.169056 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.169476 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.177023 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-f5n2g" event={"ID":"d1d05966-d105-4c45-bfa2-e66a67882baf","Type":"ContainerStarted","Data":"35a8557dbdf27af515a9f6e30d7e692b787c70cfb3a58967964d4a12a9ccea12"} Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.177082 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-f5n2g" event={"ID":"d1d05966-d105-4c45-bfa2-e66a67882baf","Type":"ContainerStarted","Data":"e8ba5226619f2620d146678c29fc7042c70e2dbf434d91f6483bbbba2262dc44"} Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.181251 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"431139568ddf08b4ca14a4ec5fc2cb3a3621c6246164b8a0bb6e6243766d7fc9"} Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.185246 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.193436 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnwn4\" (UniqueName: \"kubernetes.io/projected/5b9d4194-45b6-445b-b882-1982630eaecf-kube-api-access-wnwn4\") pod \"machine-config-controller-84d6567774-bbpqw\" (UID: \"5b9d4194-45b6-445b-b882-1982630eaecf\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bbpqw" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.194474 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m86j\" (UniqueName: \"kubernetes.io/projected/1df1b2d7-6ec6-4b8c-95d6-d08bdf698974-kube-api-access-4m86j\") pod \"olm-operator-6b444d44fb-9s5ck\" (UID: \"1df1b2d7-6ec6-4b8c-95d6-d08bdf698974\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9s5ck" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.216375 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmxp9\" (UniqueName: \"kubernetes.io/projected/2390b3ae-cb4b-4e89-854e-af5c0cf04e1d-kube-api-access-zmxp9\") pod \"service-ca-9c57cc56f-65h6c\" (UID: \"2390b3ae-cb4b-4e89-854e-af5c0cf04e1d\") " pod="openshift-service-ca/service-ca-9c57cc56f-65h6c" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.228581 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-c9v5l" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.236052 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-khvnh" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.238967 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5thcd" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.242443 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sqnw\" (UniqueName: \"kubernetes.io/projected/aed96ecb-ad63-425d-b8f4-bd55aeb752dd-kube-api-access-4sqnw\") pod \"kube-storage-version-migrator-operator-b67b599dd-2q58n\" (UID: \"aed96ecb-ad63-425d-b8f4-bd55aeb752dd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2q58n" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.253595 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qc5tq"] Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.255206 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b4928d31-a7c1-47f0-a726-712de834e034-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-sd5f5\" (UID: \"b4928d31-a7c1-47f0-a726-712de834e034\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd5f5" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.276230 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-bp44p" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.280362 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/14c95e94-4284-423b-b1cf-8df8ef663d65-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qs7dj\" (UID: \"14c95e94-4284-423b-b1cf-8df8ef663d65\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qs7dj" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.297514 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-6m4ch" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.305037 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd5f5" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.310480 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-h5tfm"] Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.312035 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qvm82"] Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.314541 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx4x8\" (UniqueName: \"kubernetes.io/projected/ffb878ee-7cd7-4830-bb19-6b2a5e6bd1e0-kube-api-access-kx4x8\") pod \"multus-admission-controller-857f4d67dd-lxfpb\" (UID: \"ffb878ee-7cd7-4830-bb19-6b2a5e6bd1e0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfpb" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.316549 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.318257 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.321426 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b55rn\" (UniqueName: \"kubernetes.io/projected/b999de84-1fbb-4184-a136-ce68641fc110-kube-api-access-b55rn\") pod \"service-ca-operator-777779d784-f8hcm\" (UID: \"b999de84-1fbb-4184-a136-ce68641fc110\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8hcm" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.327334 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2q58n" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.334605 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfpb" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.340044 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 21 08:29:44 crc kubenswrapper[4696]: W0321 08:29:44.342618 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad264986_2840_4a2a_bee9_9ea192cdbe86.slice/crio-72d259d89c5e0442ce57bae353986386d0a6994c4d0ba775049bd8b1bbb495f0 WatchSource:0}: Error finding container 72d259d89c5e0442ce57bae353986386d0a6994c4d0ba775049bd8b1bbb495f0: Status 404 returned error can't find the container with id 72d259d89c5e0442ce57bae353986386d0a6994c4d0ba775049bd8b1bbb495f0 Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.351263 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-59mfl" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.360411 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qs7dj" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.400407 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/96b62643-30a8-47d5-8fe3-20eadf06e0f7-registry-certificates\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.400511 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89b55c04-e4f6-4cbe-8e92-5cf8726c1486-config\") pod \"kube-apiserver-operator-766d6c64bb-r656w\" (UID: \"89b55c04-e4f6-4cbe-8e92-5cf8726c1486\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r656w" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.400551 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/45a093d5-91a7-44bf-8072-fc24898b79e8-metrics-tls\") pod \"ingress-operator-5b745b69d9-ntmrq\" (UID: \"45a093d5-91a7-44bf-8072-fc24898b79e8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ntmrq" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.400603 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/96b62643-30a8-47d5-8fe3-20eadf06e0f7-installation-pull-secrets\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.400741 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/45a093d5-91a7-44bf-8072-fc24898b79e8-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ntmrq\" (UID: \"45a093d5-91a7-44bf-8072-fc24898b79e8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ntmrq" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.400797 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdk58\" (UniqueName: \"kubernetes.io/projected/973fa2b5-c6ce-498f-9b2b-311e8dd34057-kube-api-access-sdk58\") pod \"marketplace-operator-79b997595-5vrbx\" (UID: \"973fa2b5-c6ce-498f-9b2b-311e8dd34057\") " pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.400856 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdgpq\" (UniqueName: \"kubernetes.io/projected/6769f68b-1f58-48e4-8076-14fcdb9c8036-kube-api-access-cdgpq\") pod \"control-plane-machine-set-operator-78cbb6b69f-5mlz9\" (UID: \"6769f68b-1f58-48e4-8076-14fcdb9c8036\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5mlz9" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.400905 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7v7t\" (UniqueName: \"kubernetes.io/projected/231fd5a3-3560-4a89-a5be-59880089a8fe-kube-api-access-h7v7t\") pod \"router-default-5444994796-sdgwl\" (UID: \"231fd5a3-3560-4a89-a5be-59880089a8fe\") " pod="openshift-ingress/router-default-5444994796-sdgwl" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.400943 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.400961 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/973fa2b5-c6ce-498f-9b2b-311e8dd34057-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5vrbx\" (UID: \"973fa2b5-c6ce-498f-9b2b-311e8dd34057\") " pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.401002 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8a77c875-c8d2-40e5-aabc-4f7ffcd29693-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-qm5pb\" (UID: \"8a77c875-c8d2-40e5-aabc-4f7ffcd29693\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qm5pb" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.401020 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/231fd5a3-3560-4a89-a5be-59880089a8fe-metrics-certs\") pod \"router-default-5444994796-sdgwl\" (UID: \"231fd5a3-3560-4a89-a5be-59880089a8fe\") " pod="openshift-ingress/router-default-5444994796-sdgwl" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.401037 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pj4n5\" (UniqueName: \"kubernetes.io/projected/96b62643-30a8-47d5-8fe3-20eadf06e0f7-kube-api-access-pj4n5\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.401107 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/528125c4-7d83-4707-ae0d-da40498497d0-node-bootstrap-token\") pod \"machine-config-server-44xzn\" (UID: \"528125c4-7d83-4707-ae0d-da40498497d0\") " pod="openshift-machine-config-operator/machine-config-server-44xzn" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.401158 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/96b62643-30a8-47d5-8fe3-20eadf06e0f7-bound-sa-token\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.401176 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2zdt\" (UniqueName: \"kubernetes.io/projected/528125c4-7d83-4707-ae0d-da40498497d0-kube-api-access-q2zdt\") pod \"machine-config-server-44xzn\" (UID: \"528125c4-7d83-4707-ae0d-da40498497d0\") " pod="openshift-machine-config-operator/machine-config-server-44xzn" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.401227 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk24q\" (UniqueName: \"kubernetes.io/projected/f3f5d127-addf-4f3a-a9f0-fcb502db57c0-kube-api-access-wk24q\") pod \"downloads-7954f5f757-jdfkn\" (UID: \"f3f5d127-addf-4f3a-a9f0-fcb502db57c0\") " pod="openshift-console/downloads-7954f5f757-jdfkn" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.401247 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/96b62643-30a8-47d5-8fe3-20eadf06e0f7-ca-trust-extracted\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.401263 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/96b62643-30a8-47d5-8fe3-20eadf06e0f7-registry-tls\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.401278 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/973fa2b5-c6ce-498f-9b2b-311e8dd34057-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5vrbx\" (UID: \"973fa2b5-c6ce-498f-9b2b-311e8dd34057\") " pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.401296 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96b62643-30a8-47d5-8fe3-20eadf06e0f7-trusted-ca\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.401313 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/528125c4-7d83-4707-ae0d-da40498497d0-certs\") pod \"machine-config-server-44xzn\" (UID: \"528125c4-7d83-4707-ae0d-da40498497d0\") " pod="openshift-machine-config-operator/machine-config-server-44xzn" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.401355 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/231fd5a3-3560-4a89-a5be-59880089a8fe-default-certificate\") pod \"router-default-5444994796-sdgwl\" (UID: \"231fd5a3-3560-4a89-a5be-59880089a8fe\") " pod="openshift-ingress/router-default-5444994796-sdgwl" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.401442 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89b55c04-e4f6-4cbe-8e92-5cf8726c1486-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-r656w\" (UID: \"89b55c04-e4f6-4cbe-8e92-5cf8726c1486\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r656w" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.401469 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k68tg\" (UniqueName: \"kubernetes.io/projected/8a77c875-c8d2-40e5-aabc-4f7ffcd29693-kube-api-access-k68tg\") pod \"package-server-manager-789f6589d5-qm5pb\" (UID: \"8a77c875-c8d2-40e5-aabc-4f7ffcd29693\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qm5pb" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.401487 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/45a093d5-91a7-44bf-8072-fc24898b79e8-trusted-ca\") pod \"ingress-operator-5b745b69d9-ntmrq\" (UID: \"45a093d5-91a7-44bf-8072-fc24898b79e8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ntmrq" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.401506 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvfmd\" (UniqueName: \"kubernetes.io/projected/45a093d5-91a7-44bf-8072-fc24898b79e8-kube-api-access-hvfmd\") pod \"ingress-operator-5b745b69d9-ntmrq\" (UID: \"45a093d5-91a7-44bf-8072-fc24898b79e8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ntmrq" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.401531 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/231fd5a3-3560-4a89-a5be-59880089a8fe-service-ca-bundle\") pod \"router-default-5444994796-sdgwl\" (UID: \"231fd5a3-3560-4a89-a5be-59880089a8fe\") " pod="openshift-ingress/router-default-5444994796-sdgwl" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.401584 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6769f68b-1f58-48e4-8076-14fcdb9c8036-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5mlz9\" (UID: \"6769f68b-1f58-48e4-8076-14fcdb9c8036\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5mlz9" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.401619 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/89b55c04-e4f6-4cbe-8e92-5cf8726c1486-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-r656w\" (UID: \"89b55c04-e4f6-4cbe-8e92-5cf8726c1486\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r656w" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.401636 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/231fd5a3-3560-4a89-a5be-59880089a8fe-stats-auth\") pod \"router-default-5444994796-sdgwl\" (UID: \"231fd5a3-3560-4a89-a5be-59880089a8fe\") " pod="openshift-ingress/router-default-5444994796-sdgwl" Mar 21 08:29:44 crc kubenswrapper[4696]: E0321 08:29:44.404066 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:44.904043097 +0000 UTC m=+119.024924010 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.411619 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bbpqw" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.417908 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.437109 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.444912 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-65h6c" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.445222 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8hcm" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.451696 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9s5ck" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.466104 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.494969 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-c9v5l"] Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.502696 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.502999 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k68tg\" (UniqueName: \"kubernetes.io/projected/8a77c875-c8d2-40e5-aabc-4f7ffcd29693-kube-api-access-k68tg\") pod \"package-server-manager-789f6589d5-qm5pb\" (UID: \"8a77c875-c8d2-40e5-aabc-4f7ffcd29693\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qm5pb" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.503060 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/45a093d5-91a7-44bf-8072-fc24898b79e8-trusted-ca\") pod \"ingress-operator-5b745b69d9-ntmrq\" (UID: \"45a093d5-91a7-44bf-8072-fc24898b79e8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ntmrq" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.503123 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvfmd\" (UniqueName: \"kubernetes.io/projected/45a093d5-91a7-44bf-8072-fc24898b79e8-kube-api-access-hvfmd\") pod \"ingress-operator-5b745b69d9-ntmrq\" (UID: \"45a093d5-91a7-44bf-8072-fc24898b79e8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ntmrq" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.503186 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8425n\" (UniqueName: \"kubernetes.io/projected/6ec15f5e-2f88-4f9c-b581-d6ec874f6337-kube-api-access-8425n\") pod \"ingress-canary-4tb7m\" (UID: \"6ec15f5e-2f88-4f9c-b581-d6ec874f6337\") " pod="openshift-ingress-canary/ingress-canary-4tb7m" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.503213 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/231fd5a3-3560-4a89-a5be-59880089a8fe-service-ca-bundle\") pod \"router-default-5444994796-sdgwl\" (UID: \"231fd5a3-3560-4a89-a5be-59880089a8fe\") " pod="openshift-ingress/router-default-5444994796-sdgwl" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.503235 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcpl7\" (UniqueName: \"kubernetes.io/projected/5f37526d-ea7c-4607-920a-ca63aa96184c-kube-api-access-vcpl7\") pod \"cni-sysctl-allowlist-ds-wkbtr\" (UID: \"5f37526d-ea7c-4607-920a-ca63aa96184c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.503276 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6769f68b-1f58-48e4-8076-14fcdb9c8036-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5mlz9\" (UID: \"6769f68b-1f58-48e4-8076-14fcdb9c8036\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5mlz9" Mar 21 08:29:44 crc kubenswrapper[4696]: E0321 08:29:44.503322 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:45.003297978 +0000 UTC m=+119.124178701 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.503385 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/89b55c04-e4f6-4cbe-8e92-5cf8726c1486-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-r656w\" (UID: \"89b55c04-e4f6-4cbe-8e92-5cf8726c1486\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r656w" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.503412 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/231fd5a3-3560-4a89-a5be-59880089a8fe-stats-auth\") pod \"router-default-5444994796-sdgwl\" (UID: \"231fd5a3-3560-4a89-a5be-59880089a8fe\") " pod="openshift-ingress/router-default-5444994796-sdgwl" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.503463 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/96b62643-30a8-47d5-8fe3-20eadf06e0f7-registry-certificates\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.503524 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89b55c04-e4f6-4cbe-8e92-5cf8726c1486-config\") pod \"kube-apiserver-operator-766d6c64bb-r656w\" (UID: \"89b55c04-e4f6-4cbe-8e92-5cf8726c1486\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r656w" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.503552 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9d4f19c5-4861-4ae6-99d0-d32e07de56e2-registration-dir\") pod \"csi-hostpathplugin-wshd6\" (UID: \"9d4f19c5-4861-4ae6-99d0-d32e07de56e2\") " pod="hostpath-provisioner/csi-hostpathplugin-wshd6" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.503612 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/45a093d5-91a7-44bf-8072-fc24898b79e8-metrics-tls\") pod \"ingress-operator-5b745b69d9-ntmrq\" (UID: \"45a093d5-91a7-44bf-8072-fc24898b79e8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ntmrq" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.503640 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5f37526d-ea7c-4607-920a-ca63aa96184c-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-wkbtr\" (UID: \"5f37526d-ea7c-4607-920a-ca63aa96184c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.503660 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzddt\" (UniqueName: \"kubernetes.io/projected/9d4f19c5-4861-4ae6-99d0-d32e07de56e2-kube-api-access-pzddt\") pod \"csi-hostpathplugin-wshd6\" (UID: \"9d4f19c5-4861-4ae6-99d0-d32e07de56e2\") " pod="hostpath-provisioner/csi-hostpathplugin-wshd6" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.503723 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/96b62643-30a8-47d5-8fe3-20eadf06e0f7-installation-pull-secrets\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.504844 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6ec15f5e-2f88-4f9c-b581-d6ec874f6337-cert\") pod \"ingress-canary-4tb7m\" (UID: \"6ec15f5e-2f88-4f9c-b581-d6ec874f6337\") " pod="openshift-ingress-canary/ingress-canary-4tb7m" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.504939 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/45a093d5-91a7-44bf-8072-fc24898b79e8-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ntmrq\" (UID: \"45a093d5-91a7-44bf-8072-fc24898b79e8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ntmrq" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.505255 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdk58\" (UniqueName: \"kubernetes.io/projected/973fa2b5-c6ce-498f-9b2b-311e8dd34057-kube-api-access-sdk58\") pod \"marketplace-operator-79b997595-5vrbx\" (UID: \"973fa2b5-c6ce-498f-9b2b-311e8dd34057\") " pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.505425 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdgpq\" (UniqueName: \"kubernetes.io/projected/6769f68b-1f58-48e4-8076-14fcdb9c8036-kube-api-access-cdgpq\") pod \"control-plane-machine-set-operator-78cbb6b69f-5mlz9\" (UID: \"6769f68b-1f58-48e4-8076-14fcdb9c8036\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5mlz9" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.505623 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7v7t\" (UniqueName: \"kubernetes.io/projected/231fd5a3-3560-4a89-a5be-59880089a8fe-kube-api-access-h7v7t\") pod \"router-default-5444994796-sdgwl\" (UID: \"231fd5a3-3560-4a89-a5be-59880089a8fe\") " pod="openshift-ingress/router-default-5444994796-sdgwl" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.505651 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxlpb\" (UniqueName: \"kubernetes.io/projected/367f9b3a-40fe-414e-bc06-949205afe2b8-kube-api-access-kxlpb\") pod \"dns-default-chtnc\" (UID: \"367f9b3a-40fe-414e-bc06-949205afe2b8\") " pod="openshift-dns/dns-default-chtnc" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.505884 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5f37526d-ea7c-4607-920a-ca63aa96184c-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-wkbtr\" (UID: \"5f37526d-ea7c-4607-920a-ca63aa96184c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.506056 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.506118 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/973fa2b5-c6ce-498f-9b2b-311e8dd34057-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5vrbx\" (UID: \"973fa2b5-c6ce-498f-9b2b-311e8dd34057\") " pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.506140 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/231fd5a3-3560-4a89-a5be-59880089a8fe-metrics-certs\") pod \"router-default-5444994796-sdgwl\" (UID: \"231fd5a3-3560-4a89-a5be-59880089a8fe\") " pod="openshift-ingress/router-default-5444994796-sdgwl" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.506177 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/367f9b3a-40fe-414e-bc06-949205afe2b8-metrics-tls\") pod \"dns-default-chtnc\" (UID: \"367f9b3a-40fe-414e-bc06-949205afe2b8\") " pod="openshift-dns/dns-default-chtnc" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.506201 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8a77c875-c8d2-40e5-aabc-4f7ffcd29693-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-qm5pb\" (UID: \"8a77c875-c8d2-40e5-aabc-4f7ffcd29693\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qm5pb" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.506231 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pj4n5\" (UniqueName: \"kubernetes.io/projected/96b62643-30a8-47d5-8fe3-20eadf06e0f7-kube-api-access-pj4n5\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.506307 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/528125c4-7d83-4707-ae0d-da40498497d0-node-bootstrap-token\") pod \"machine-config-server-44xzn\" (UID: \"528125c4-7d83-4707-ae0d-da40498497d0\") " pod="openshift-machine-config-operator/machine-config-server-44xzn" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.506366 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/96b62643-30a8-47d5-8fe3-20eadf06e0f7-bound-sa-token\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.506396 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2zdt\" (UniqueName: \"kubernetes.io/projected/528125c4-7d83-4707-ae0d-da40498497d0-kube-api-access-q2zdt\") pod \"machine-config-server-44xzn\" (UID: \"528125c4-7d83-4707-ae0d-da40498497d0\") " pod="openshift-machine-config-operator/machine-config-server-44xzn" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.506458 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk24q\" (UniqueName: \"kubernetes.io/projected/f3f5d127-addf-4f3a-a9f0-fcb502db57c0-kube-api-access-wk24q\") pod \"downloads-7954f5f757-jdfkn\" (UID: \"f3f5d127-addf-4f3a-a9f0-fcb502db57c0\") " pod="openshift-console/downloads-7954f5f757-jdfkn" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.506573 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/96b62643-30a8-47d5-8fe3-20eadf06e0f7-ca-trust-extracted\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.506741 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/367f9b3a-40fe-414e-bc06-949205afe2b8-config-volume\") pod \"dns-default-chtnc\" (UID: \"367f9b3a-40fe-414e-bc06-949205afe2b8\") " pod="openshift-dns/dns-default-chtnc" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.506973 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/96b62643-30a8-47d5-8fe3-20eadf06e0f7-ca-trust-extracted\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.506989 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9d4f19c5-4861-4ae6-99d0-d32e07de56e2-socket-dir\") pod \"csi-hostpathplugin-wshd6\" (UID: \"9d4f19c5-4861-4ae6-99d0-d32e07de56e2\") " pod="hostpath-provisioner/csi-hostpathplugin-wshd6" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.507033 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/973fa2b5-c6ce-498f-9b2b-311e8dd34057-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5vrbx\" (UID: \"973fa2b5-c6ce-498f-9b2b-311e8dd34057\") " pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.508302 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/96b62643-30a8-47d5-8fe3-20eadf06e0f7-registry-tls\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.508367 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96b62643-30a8-47d5-8fe3-20eadf06e0f7-trusted-ca\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.508468 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9d4f19c5-4861-4ae6-99d0-d32e07de56e2-mountpoint-dir\") pod \"csi-hostpathplugin-wshd6\" (UID: \"9d4f19c5-4861-4ae6-99d0-d32e07de56e2\") " pod="hostpath-provisioner/csi-hostpathplugin-wshd6" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.508651 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/528125c4-7d83-4707-ae0d-da40498497d0-certs\") pod \"machine-config-server-44xzn\" (UID: \"528125c4-7d83-4707-ae0d-da40498497d0\") " pod="openshift-machine-config-operator/machine-config-server-44xzn" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.508707 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9d4f19c5-4861-4ae6-99d0-d32e07de56e2-plugins-dir\") pod \"csi-hostpathplugin-wshd6\" (UID: \"9d4f19c5-4861-4ae6-99d0-d32e07de56e2\") " pod="hostpath-provisioner/csi-hostpathplugin-wshd6" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.508730 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/231fd5a3-3560-4a89-a5be-59880089a8fe-default-certificate\") pod \"router-default-5444994796-sdgwl\" (UID: \"231fd5a3-3560-4a89-a5be-59880089a8fe\") " pod="openshift-ingress/router-default-5444994796-sdgwl" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.508787 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/5f37526d-ea7c-4607-920a-ca63aa96184c-ready\") pod \"cni-sysctl-allowlist-ds-wkbtr\" (UID: \"5f37526d-ea7c-4607-920a-ca63aa96184c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.509032 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9d4f19c5-4861-4ae6-99d0-d32e07de56e2-csi-data-dir\") pod \"csi-hostpathplugin-wshd6\" (UID: \"9d4f19c5-4861-4ae6-99d0-d32e07de56e2\") " pod="hostpath-provisioner/csi-hostpathplugin-wshd6" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.509132 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89b55c04-e4f6-4cbe-8e92-5cf8726c1486-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-r656w\" (UID: \"89b55c04-e4f6-4cbe-8e92-5cf8726c1486\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r656w" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.511663 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/45a093d5-91a7-44bf-8072-fc24898b79e8-trusted-ca\") pod \"ingress-operator-5b745b69d9-ntmrq\" (UID: \"45a093d5-91a7-44bf-8072-fc24898b79e8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ntmrq" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.518393 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/231fd5a3-3560-4a89-a5be-59880089a8fe-service-ca-bundle\") pod \"router-default-5444994796-sdgwl\" (UID: \"231fd5a3-3560-4a89-a5be-59880089a8fe\") " pod="openshift-ingress/router-default-5444994796-sdgwl" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.524660 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/973fa2b5-c6ce-498f-9b2b-311e8dd34057-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5vrbx\" (UID: \"973fa2b5-c6ce-498f-9b2b-311e8dd34057\") " pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.525196 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6769f68b-1f58-48e4-8076-14fcdb9c8036-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5mlz9\" (UID: \"6769f68b-1f58-48e4-8076-14fcdb9c8036\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5mlz9" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.525483 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/231fd5a3-3560-4a89-a5be-59880089a8fe-stats-auth\") pod \"router-default-5444994796-sdgwl\" (UID: \"231fd5a3-3560-4a89-a5be-59880089a8fe\") " pod="openshift-ingress/router-default-5444994796-sdgwl" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.527486 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/96b62643-30a8-47d5-8fe3-20eadf06e0f7-registry-certificates\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: E0321 08:29:44.527704 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:45.027687073 +0000 UTC m=+119.148567796 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.528566 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96b62643-30a8-47d5-8fe3-20eadf06e0f7-trusted-ca\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.529343 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89b55c04-e4f6-4cbe-8e92-5cf8726c1486-config\") pod \"kube-apiserver-operator-766d6c64bb-r656w\" (UID: \"89b55c04-e4f6-4cbe-8e92-5cf8726c1486\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r656w" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.530416 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/973fa2b5-c6ce-498f-9b2b-311e8dd34057-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5vrbx\" (UID: \"973fa2b5-c6ce-498f-9b2b-311e8dd34057\") " pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.532518 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/45a093d5-91a7-44bf-8072-fc24898b79e8-metrics-tls\") pod \"ingress-operator-5b745b69d9-ntmrq\" (UID: \"45a093d5-91a7-44bf-8072-fc24898b79e8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ntmrq" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.533292 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/96b62643-30a8-47d5-8fe3-20eadf06e0f7-installation-pull-secrets\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.535382 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89b55c04-e4f6-4cbe-8e92-5cf8726c1486-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-r656w\" (UID: \"89b55c04-e4f6-4cbe-8e92-5cf8726c1486\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r656w" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.569959 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/528125c4-7d83-4707-ae0d-da40498497d0-certs\") pod \"machine-config-server-44xzn\" (UID: \"528125c4-7d83-4707-ae0d-da40498497d0\") " pod="openshift-machine-config-operator/machine-config-server-44xzn" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.571089 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/231fd5a3-3560-4a89-a5be-59880089a8fe-default-certificate\") pod \"router-default-5444994796-sdgwl\" (UID: \"231fd5a3-3560-4a89-a5be-59880089a8fe\") " pod="openshift-ingress/router-default-5444994796-sdgwl" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.571200 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8a77c875-c8d2-40e5-aabc-4f7ffcd29693-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-qm5pb\" (UID: \"8a77c875-c8d2-40e5-aabc-4f7ffcd29693\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qm5pb" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.571914 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/528125c4-7d83-4707-ae0d-da40498497d0-node-bootstrap-token\") pod \"machine-config-server-44xzn\" (UID: \"528125c4-7d83-4707-ae0d-da40498497d0\") " pod="openshift-machine-config-operator/machine-config-server-44xzn" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.572473 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/96b62643-30a8-47d5-8fe3-20eadf06e0f7-registry-tls\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.572705 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/231fd5a3-3560-4a89-a5be-59880089a8fe-metrics-certs\") pod \"router-default-5444994796-sdgwl\" (UID: \"231fd5a3-3560-4a89-a5be-59880089a8fe\") " pod="openshift-ingress/router-default-5444994796-sdgwl" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.576382 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvfmd\" (UniqueName: \"kubernetes.io/projected/45a093d5-91a7-44bf-8072-fc24898b79e8-kube-api-access-hvfmd\") pod \"ingress-operator-5b745b69d9-ntmrq\" (UID: \"45a093d5-91a7-44bf-8072-fc24898b79e8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ntmrq" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.578994 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k68tg\" (UniqueName: \"kubernetes.io/projected/8a77c875-c8d2-40e5-aabc-4f7ffcd29693-kube-api-access-k68tg\") pod \"package-server-manager-789f6589d5-qm5pb\" (UID: \"8a77c875-c8d2-40e5-aabc-4f7ffcd29693\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qm5pb" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.636613 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/89b55c04-e4f6-4cbe-8e92-5cf8726c1486-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-r656w\" (UID: \"89b55c04-e4f6-4cbe-8e92-5cf8726c1486\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r656w" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.639374 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:44 crc kubenswrapper[4696]: E0321 08:29:44.639745 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:45.139725888 +0000 UTC m=+119.260606601 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.639864 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6ec15f5e-2f88-4f9c-b581-d6ec874f6337-cert\") pod \"ingress-canary-4tb7m\" (UID: \"6ec15f5e-2f88-4f9c-b581-d6ec874f6337\") " pod="openshift-ingress-canary/ingress-canary-4tb7m" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.639932 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxlpb\" (UniqueName: \"kubernetes.io/projected/367f9b3a-40fe-414e-bc06-949205afe2b8-kube-api-access-kxlpb\") pod \"dns-default-chtnc\" (UID: \"367f9b3a-40fe-414e-bc06-949205afe2b8\") " pod="openshift-dns/dns-default-chtnc" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.639981 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5f37526d-ea7c-4607-920a-ca63aa96184c-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-wkbtr\" (UID: \"5f37526d-ea7c-4607-920a-ca63aa96184c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.640011 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.640038 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/367f9b3a-40fe-414e-bc06-949205afe2b8-metrics-tls\") pod \"dns-default-chtnc\" (UID: \"367f9b3a-40fe-414e-bc06-949205afe2b8\") " pod="openshift-dns/dns-default-chtnc" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.640099 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9d4f19c5-4861-4ae6-99d0-d32e07de56e2-socket-dir\") pod \"csi-hostpathplugin-wshd6\" (UID: \"9d4f19c5-4861-4ae6-99d0-d32e07de56e2\") " pod="hostpath-provisioner/csi-hostpathplugin-wshd6" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.640120 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/367f9b3a-40fe-414e-bc06-949205afe2b8-config-volume\") pod \"dns-default-chtnc\" (UID: \"367f9b3a-40fe-414e-bc06-949205afe2b8\") " pod="openshift-dns/dns-default-chtnc" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.640145 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9d4f19c5-4861-4ae6-99d0-d32e07de56e2-mountpoint-dir\") pod \"csi-hostpathplugin-wshd6\" (UID: \"9d4f19c5-4861-4ae6-99d0-d32e07de56e2\") " pod="hostpath-provisioner/csi-hostpathplugin-wshd6" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.640168 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9d4f19c5-4861-4ae6-99d0-d32e07de56e2-plugins-dir\") pod \"csi-hostpathplugin-wshd6\" (UID: \"9d4f19c5-4861-4ae6-99d0-d32e07de56e2\") " pod="hostpath-provisioner/csi-hostpathplugin-wshd6" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.640191 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/5f37526d-ea7c-4607-920a-ca63aa96184c-ready\") pod \"cni-sysctl-allowlist-ds-wkbtr\" (UID: \"5f37526d-ea7c-4607-920a-ca63aa96184c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.640223 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9d4f19c5-4861-4ae6-99d0-d32e07de56e2-csi-data-dir\") pod \"csi-hostpathplugin-wshd6\" (UID: \"9d4f19c5-4861-4ae6-99d0-d32e07de56e2\") " pod="hostpath-provisioner/csi-hostpathplugin-wshd6" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.640260 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8425n\" (UniqueName: \"kubernetes.io/projected/6ec15f5e-2f88-4f9c-b581-d6ec874f6337-kube-api-access-8425n\") pod \"ingress-canary-4tb7m\" (UID: \"6ec15f5e-2f88-4f9c-b581-d6ec874f6337\") " pod="openshift-ingress-canary/ingress-canary-4tb7m" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.640287 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcpl7\" (UniqueName: \"kubernetes.io/projected/5f37526d-ea7c-4607-920a-ca63aa96184c-kube-api-access-vcpl7\") pod \"cni-sysctl-allowlist-ds-wkbtr\" (UID: \"5f37526d-ea7c-4607-920a-ca63aa96184c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.640335 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9d4f19c5-4861-4ae6-99d0-d32e07de56e2-registration-dir\") pod \"csi-hostpathplugin-wshd6\" (UID: \"9d4f19c5-4861-4ae6-99d0-d32e07de56e2\") " pod="hostpath-provisioner/csi-hostpathplugin-wshd6" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.640361 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5f37526d-ea7c-4607-920a-ca63aa96184c-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-wkbtr\" (UID: \"5f37526d-ea7c-4607-920a-ca63aa96184c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.640385 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzddt\" (UniqueName: \"kubernetes.io/projected/9d4f19c5-4861-4ae6-99d0-d32e07de56e2-kube-api-access-pzddt\") pod \"csi-hostpathplugin-wshd6\" (UID: \"9d4f19c5-4861-4ae6-99d0-d32e07de56e2\") " pod="hostpath-provisioner/csi-hostpathplugin-wshd6" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.644025 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdgpq\" (UniqueName: \"kubernetes.io/projected/6769f68b-1f58-48e4-8076-14fcdb9c8036-kube-api-access-cdgpq\") pod \"control-plane-machine-set-operator-78cbb6b69f-5mlz9\" (UID: \"6769f68b-1f58-48e4-8076-14fcdb9c8036\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5mlz9" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.644747 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9d4f19c5-4861-4ae6-99d0-d32e07de56e2-mountpoint-dir\") pod \"csi-hostpathplugin-wshd6\" (UID: \"9d4f19c5-4861-4ae6-99d0-d32e07de56e2\") " pod="hostpath-provisioner/csi-hostpathplugin-wshd6" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.645045 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9d4f19c5-4861-4ae6-99d0-d32e07de56e2-csi-data-dir\") pod \"csi-hostpathplugin-wshd6\" (UID: \"9d4f19c5-4861-4ae6-99d0-d32e07de56e2\") " pod="hostpath-provisioner/csi-hostpathplugin-wshd6" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.645575 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/367f9b3a-40fe-414e-bc06-949205afe2b8-config-volume\") pod \"dns-default-chtnc\" (UID: \"367f9b3a-40fe-414e-bc06-949205afe2b8\") " pod="openshift-dns/dns-default-chtnc" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.646149 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/5f37526d-ea7c-4607-920a-ca63aa96184c-ready\") pod \"cni-sysctl-allowlist-ds-wkbtr\" (UID: \"5f37526d-ea7c-4607-920a-ca63aa96184c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.646679 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9d4f19c5-4861-4ae6-99d0-d32e07de56e2-plugins-dir\") pod \"csi-hostpathplugin-wshd6\" (UID: \"9d4f19c5-4861-4ae6-99d0-d32e07de56e2\") " pod="hostpath-provisioner/csi-hostpathplugin-wshd6" Mar 21 08:29:44 crc kubenswrapper[4696]: E0321 08:29:44.649192 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:45.147796708 +0000 UTC m=+119.268677511 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.650414 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5f37526d-ea7c-4607-920a-ca63aa96184c-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-wkbtr\" (UID: \"5f37526d-ea7c-4607-920a-ca63aa96184c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.650473 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9d4f19c5-4861-4ae6-99d0-d32e07de56e2-socket-dir\") pod \"csi-hostpathplugin-wshd6\" (UID: \"9d4f19c5-4861-4ae6-99d0-d32e07de56e2\") " pod="hostpath-provisioner/csi-hostpathplugin-wshd6" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.650515 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9d4f19c5-4861-4ae6-99d0-d32e07de56e2-registration-dir\") pod \"csi-hostpathplugin-wshd6\" (UID: \"9d4f19c5-4861-4ae6-99d0-d32e07de56e2\") " pod="hostpath-provisioner/csi-hostpathplugin-wshd6" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.652262 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5f37526d-ea7c-4607-920a-ca63aa96184c-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-wkbtr\" (UID: \"5f37526d-ea7c-4607-920a-ca63aa96184c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.657648 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5thcd"] Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.661652 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdk58\" (UniqueName: \"kubernetes.io/projected/973fa2b5-c6ce-498f-9b2b-311e8dd34057-kube-api-access-sdk58\") pod \"marketplace-operator-79b997595-5vrbx\" (UID: \"973fa2b5-c6ce-498f-9b2b-311e8dd34057\") " pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.664762 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6ec15f5e-2f88-4f9c-b581-d6ec874f6337-cert\") pod \"ingress-canary-4tb7m\" (UID: \"6ec15f5e-2f88-4f9c-b581-d6ec874f6337\") " pod="openshift-ingress-canary/ingress-canary-4tb7m" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.666433 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5mlz9" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.666637 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/367f9b3a-40fe-414e-bc06-949205afe2b8-metrics-tls\") pod \"dns-default-chtnc\" (UID: \"367f9b3a-40fe-414e-bc06-949205afe2b8\") " pod="openshift-dns/dns-default-chtnc" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.674411 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/45a093d5-91a7-44bf-8072-fc24898b79e8-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ntmrq\" (UID: \"45a093d5-91a7-44bf-8072-fc24898b79e8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ntmrq" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.675188 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7v7t\" (UniqueName: \"kubernetes.io/projected/231fd5a3-3560-4a89-a5be-59880089a8fe-kube-api-access-h7v7t\") pod \"router-default-5444994796-sdgwl\" (UID: \"231fd5a3-3560-4a89-a5be-59880089a8fe\") " pod="openshift-ingress/router-default-5444994796-sdgwl" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.675921 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ntmrq" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.686626 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.694144 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r656w" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.698771 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-sdgwl" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.699915 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/96b62643-30a8-47d5-8fe3-20eadf06e0f7-bound-sa-token\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.718254 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pj4n5\" (UniqueName: \"kubernetes.io/projected/96b62643-30a8-47d5-8fe3-20eadf06e0f7-kube-api-access-pj4n5\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.742717 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:44 crc kubenswrapper[4696]: E0321 08:29:44.743218 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:45.24320124 +0000 UTC m=+119.364081943 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.743858 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r"] Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.755066 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk24q\" (UniqueName: \"kubernetes.io/projected/f3f5d127-addf-4f3a-a9f0-fcb502db57c0-kube-api-access-wk24q\") pod \"downloads-7954f5f757-jdfkn\" (UID: \"f3f5d127-addf-4f3a-a9f0-fcb502db57c0\") " pod="openshift-console/downloads-7954f5f757-jdfkn" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.761223 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2zdt\" (UniqueName: \"kubernetes.io/projected/528125c4-7d83-4707-ae0d-da40498497d0-kube-api-access-q2zdt\") pod \"machine-config-server-44xzn\" (UID: \"528125c4-7d83-4707-ae0d-da40498497d0\") " pod="openshift-machine-config-operator/machine-config-server-44xzn" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.787962 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzddt\" (UniqueName: \"kubernetes.io/projected/9d4f19c5-4861-4ae6-99d0-d32e07de56e2-kube-api-access-pzddt\") pod \"csi-hostpathplugin-wshd6\" (UID: \"9d4f19c5-4861-4ae6-99d0-d32e07de56e2\") " pod="hostpath-provisioner/csi-hostpathplugin-wshd6" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.789426 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qm5pb" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.804402 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxlpb\" (UniqueName: \"kubernetes.io/projected/367f9b3a-40fe-414e-bc06-949205afe2b8-kube-api-access-kxlpb\") pod \"dns-default-chtnc\" (UID: \"367f9b3a-40fe-414e-bc06-949205afe2b8\") " pod="openshift-dns/dns-default-chtnc" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.809373 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-44xzn" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.812194 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-wshd6" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.849446 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8425n\" (UniqueName: \"kubernetes.io/projected/6ec15f5e-2f88-4f9c-b581-d6ec874f6337-kube-api-access-8425n\") pod \"ingress-canary-4tb7m\" (UID: \"6ec15f5e-2f88-4f9c-b581-d6ec874f6337\") " pod="openshift-ingress-canary/ingress-canary-4tb7m" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.849465 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.857216 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.857463 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcpl7\" (UniqueName: \"kubernetes.io/projected/5f37526d-ea7c-4607-920a-ca63aa96184c-kube-api-access-vcpl7\") pod \"cni-sysctl-allowlist-ds-wkbtr\" (UID: \"5f37526d-ea7c-4607-920a-ca63aa96184c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" Mar 21 08:29:44 crc kubenswrapper[4696]: E0321 08:29:44.857636 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:45.357624546 +0000 UTC m=+119.478505259 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.914275 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-jdfkn" Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.958846 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:44 crc kubenswrapper[4696]: E0321 08:29:44.959062 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:45.459023954 +0000 UTC m=+119.579904667 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:44 crc kubenswrapper[4696]: I0321 08:29:44.959178 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:44 crc kubenswrapper[4696]: E0321 08:29:44.959762 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:45.459745102 +0000 UTC m=+119.580625815 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.060373 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:45 crc kubenswrapper[4696]: E0321 08:29:45.060835 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:45.560786091 +0000 UTC m=+119.681666804 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.060963 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:45 crc kubenswrapper[4696]: E0321 08:29:45.061414 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:45.561400177 +0000 UTC m=+119.682280890 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.085211 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-chtnc" Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.120515 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.121925 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-4tb7m" Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.165786 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:45 crc kubenswrapper[4696]: E0321 08:29:45.166650 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:45.666284416 +0000 UTC m=+119.787165129 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.185962 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qs7dj"] Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.190303 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd5f5"] Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.230051 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-h5tfm" event={"ID":"f62330e9-cdc5-469b-82ba-3364ce173686","Type":"ContainerStarted","Data":"7439d365291807f4de5becd7d840cf3947a1105556d8dc8a3ab4ee9552325178"} Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.233954 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-6m4ch"] Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.239049 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r" event={"ID":"44b39f4b-4213-452c-94f7-946308af1e2f","Type":"ContainerStarted","Data":"c4226490c82f33717dd4a15853e09aa67b3dcc6d9bcf5be5ad5fa8a0f829fe02"} Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.257762 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bp44p"] Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.271157 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:45 crc kubenswrapper[4696]: E0321 08:29:45.271583 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:45.771567664 +0000 UTC m=+119.892448377 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.282074 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5thcd" event={"ID":"b83c2afe-6390-4518-a5c1-c6ca14ea4629","Type":"ContainerStarted","Data":"076a32ae8d12baa8f2ee603fce4aa1984af127f50d2cd1b1ac209a07f2eb6469"} Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.304662 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-khvnh" event={"ID":"ad264986-2840-4a2a-bee9-9ea192cdbe86","Type":"ContainerStarted","Data":"72d259d89c5e0442ce57bae353986386d0a6994c4d0ba775049bd8b1bbb495f0"} Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.343418 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" event={"ID":"851ce288-d60d-4875-ae01-cc13268fec8c","Type":"ContainerStarted","Data":"f63c031a8c919e759aad83e9d49ec2a6b148c4ce571fa9a3c83bc482c55a72f5"} Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.358147 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-9thjt" event={"ID":"3239cfe6-5d28-4efc-b0b3-53699ca2a449","Type":"ContainerStarted","Data":"b412b83d91fcb537983fc45dc5a013e300730afe85cfdb52ab793924dc407062"} Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.358211 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-9thjt" event={"ID":"3239cfe6-5d28-4efc-b0b3-53699ca2a449","Type":"ContainerStarted","Data":"4afffd42ed77924033235252336ab078226669460fd371b38a89ba992f372f5c"} Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.362177 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-9thjt" Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.373188 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"00a2dbd1c960d90b852dd83ddd21cf11a359ec509982fa07d29c77ee6d144486"} Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.374347 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:45 crc kubenswrapper[4696]: E0321 08:29:45.374751 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:45.874731798 +0000 UTC m=+119.995612511 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.378590 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-xblff" event={"ID":"e24de32f-7778-4fac-815b-fa4d2a44dd2f","Type":"ContainerStarted","Data":"9c83e397edd78eb86ec4732ad8558e5d7caaefdf4a630a0a891a3344759f7050"} Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.387199 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qc5tq" event={"ID":"dc140eb2-a579-45bc-9fb4-247c254533ea","Type":"ContainerStarted","Data":"15f12575783a2a613f90c03173570c0b24d78f7fb5ca18c64e3b5c40424085d9"} Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.411684 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xt5jj" event={"ID":"68a1b8db-aec4-4681-b1f0-87761e34c518","Type":"ContainerStarted","Data":"adf620e3ba6e68a42c2cf904eaf0351792cace59c5236a9995268b77e53ef700"} Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.416570 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xt5jj" Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.420241 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2q58n"] Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.429828 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-c9v5l" event={"ID":"13b3ee79-f56b-4fff-ad26-ac72a3ef5017","Type":"ContainerStarted","Data":"75faa44ab3fa156f6f2750d00bf29a76906b6d87f5246b117e6fe4236cbb17e7"} Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.448275 4696 patch_prober.go:28] interesting pod/console-operator-58897d9998-9thjt container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.448614 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-9thjt" podUID="3239cfe6-5d28-4efc-b0b3-53699ca2a449" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.467215 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.475891 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.478748 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-lxfpb"] Mar 21 08:29:45 crc kubenswrapper[4696]: E0321 08:29:45.480676 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:45.980659554 +0000 UTC m=+120.101540267 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.580525 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:45 crc kubenswrapper[4696]: E0321 08:29:45.581968 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:46.081952649 +0000 UTC m=+120.202833362 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.685567 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:45 crc kubenswrapper[4696]: E0321 08:29:45.686296 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:46.186280703 +0000 UTC m=+120.307161416 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.746230 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-rvwvm" podStartSLOduration=78.746208222 podStartE2EDuration="1m18.746208222s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:45.542590575 +0000 UTC m=+119.663471288" watchObservedRunningTime="2026-03-21 08:29:45.746208222 +0000 UTC m=+119.867088935" Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.789251 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:45 crc kubenswrapper[4696]: E0321 08:29:45.789663 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:46.289649012 +0000 UTC m=+120.410529725 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.817548 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vnbgm" podStartSLOduration=79.817528197 podStartE2EDuration="1m19.817528197s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:45.772947707 +0000 UTC m=+119.893828410" watchObservedRunningTime="2026-03-21 08:29:45.817528197 +0000 UTC m=+119.938408910" Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.858918 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" podStartSLOduration=79.858895494 podStartE2EDuration="1m19.858895494s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:45.846158533 +0000 UTC m=+119.967039256" watchObservedRunningTime="2026-03-21 08:29:45.858895494 +0000 UTC m=+119.979776217" Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.896955 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:45 crc kubenswrapper[4696]: E0321 08:29:45.897284 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:46.397271592 +0000 UTC m=+120.518152305 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:45 crc kubenswrapper[4696]: I0321 08:29:45.998392 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:46 crc kubenswrapper[4696]: E0321 08:29:45.999141 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:46.499110421 +0000 UTC m=+120.619991134 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.053758 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ptg5t"] Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.069160 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6"] Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.073787 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-59mfl"] Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.103736 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9bbfb96a-607b-47df-83df-7c75dd94ad71-metrics-certs\") pod \"network-metrics-daemon-w95sn\" (UID: \"9bbfb96a-607b-47df-83df-7c75dd94ad71\") " pod="openshift-multus/network-metrics-daemon-w95sn" Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.103857 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:46 crc kubenswrapper[4696]: E0321 08:29:46.104255 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:46.604242096 +0000 UTC m=+120.725122809 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.121519 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx"] Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.132629 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9bbfb96a-607b-47df-83df-7c75dd94ad71-metrics-certs\") pod \"network-metrics-daemon-w95sn\" (UID: \"9bbfb96a-607b-47df-83df-7c75dd94ad71\") " pod="openshift-multus/network-metrics-daemon-w95sn" Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.136711 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-f5n2g" podStartSLOduration=80.13669444 podStartE2EDuration="1m20.13669444s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:46.12977977 +0000 UTC m=+120.250660483" watchObservedRunningTime="2026-03-21 08:29:46.13669444 +0000 UTC m=+120.257575153" Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.146898 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-w95sn" Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.153426 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l"] Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.174316 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9s5ck"] Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.206437 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:46 crc kubenswrapper[4696]: E0321 08:29:46.206951 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:46.706918267 +0000 UTC m=+120.827798980 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.262593 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5vrbx"] Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.308253 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-65h6c"] Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.309481 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:46 crc kubenswrapper[4696]: E0321 08:29:46.309918 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:46.809902886 +0000 UTC m=+120.930783599 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.321888 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-bbpqw"] Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.344765 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl"] Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.403381 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-f8hcm"] Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.410758 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:46 crc kubenswrapper[4696]: E0321 08:29:46.411294 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:46.911264173 +0000 UTC m=+121.032144886 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.512847 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:46 crc kubenswrapper[4696]: E0321 08:29:46.513223 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:47.013209365 +0000 UTC m=+121.134090078 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.518154 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qc5tq" event={"ID":"dc140eb2-a579-45bc-9fb4-247c254533ea","Type":"ContainerStarted","Data":"413fa92c4d1feb54c763d97132650c914c610f9bfcb29ed937bd8bbae422c7cd"} Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.613551 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:46 crc kubenswrapper[4696]: E0321 08:29:46.614008 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:47.113990357 +0000 UTC m=+121.234871070 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.614832 4696 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-qvm82 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.17:6443/healthz\": dial tcp 10.217.0.17:6443: connect: connection refused" start-of-body= Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.614862 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" podUID="851ce288-d60d-4875-ae01-cc13268fec8c" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.17:6443/healthz\": dial tcp 10.217.0.17:6443: connect: connection refused" Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.682602 4696 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-h5n9r container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.682656 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r" podUID="44b39f4b-4213-452c-94f7-946308af1e2f" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.716613 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.716859 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" event={"ID":"934c62b8-c7a7-4018-b5f1-7954b2baca00","Type":"ContainerStarted","Data":"5d0849eae3a99515edae959c29726dc6133161ea7ba70f9b3e2ce0405897c055"} Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.716919 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.716938 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48"] Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.716966 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" event={"ID":"851ce288-d60d-4875-ae01-cc13268fec8c","Type":"ContainerStarted","Data":"211d6c3be2967466e12208280fc9b36b1d12586d1058b439a7fc675ea195b4bc"} Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.716981 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r" Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.717201 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-59mfl" event={"ID":"bf190816-3a09-4918-ab57-d31e55919f3e","Type":"ContainerStarted","Data":"22fe50fe6f1009a2b051a692713b60bb5d5938f35562f0ceb716beb886f27454"} Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.717212 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qs7dj" event={"ID":"14c95e94-4284-423b-b1cf-8df8ef663d65","Type":"ContainerStarted","Data":"591f959ddbab66eddbf961f24ea313422043493d6cc262c93927f8a7ed8fee41"} Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.717232 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2q58n" event={"ID":"aed96ecb-ad63-425d-b8f4-bd55aeb752dd","Type":"ContainerStarted","Data":"2bfdd6d7ae96688d386bf4d0b2791f523664031860130cae0e62b0bdefafb6e3"} Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.717244 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5thcd" event={"ID":"b83c2afe-6390-4518-a5c1-c6ca14ea4629","Type":"ContainerStarted","Data":"9c7508a7d3dc4c789683a79b6410e3439753c9742818d3fa5dcd7b8fee2dc7d4"} Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.717256 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-65h6c" event={"ID":"2390b3ae-cb4b-4e89-854e-af5c0cf04e1d","Type":"ContainerStarted","Data":"4e035d2979f1a05068e2abd28108cbf39bcdbf4272e927731a2975ee374913f4"} Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.717269 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-sdgwl" event={"ID":"231fd5a3-3560-4a89-a5be-59880089a8fe","Type":"ContainerStarted","Data":"121bb9e56e1d7eda468c7f3bc3ccf591843c17a183bed79a71e63a52bf2e44ef"} Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.717282 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-6m4ch" event={"ID":"e26909f3-14e6-4f5a-9ea9-5b44ac3aa56a","Type":"ContainerStarted","Data":"76ab13291e4627f2fb308306adb516f77bebab2b5be99b141c30f4f7acbea819"} Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.718641 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r" event={"ID":"44b39f4b-4213-452c-94f7-946308af1e2f","Type":"ContainerStarted","Data":"7374c599c331e7d71decd073c2dc067100adab7ae1910de61a314c5ba4cacf12"} Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.719704 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-c9v5l" event={"ID":"13b3ee79-f56b-4fff-ad26-ac72a3ef5017","Type":"ContainerStarted","Data":"58883333ce5306df98e988e281a9eceab46d89cf0e0fcdbf3ca069af01489a7d"} Mar 21 08:29:46 crc kubenswrapper[4696]: E0321 08:29:46.721129 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:47.221108963 +0000 UTC m=+121.341989676 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.723879 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-bp44p" event={"ID":"f2a60c2e-1edd-4c96-b263-e84dec7ed4a2","Type":"ContainerStarted","Data":"78b32f451e60b758141c98a6089255c525f95210d15de03081db26a2f43b7a5e"} Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.734380 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" podStartSLOduration=79.734362728 podStartE2EDuration="1m19.734362728s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:46.732293594 +0000 UTC m=+120.853174307" watchObservedRunningTime="2026-03-21 08:29:46.734362728 +0000 UTC m=+120.855243441" Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.739946 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd5f5" event={"ID":"b4928d31-a7c1-47f0-a726-712de834e034","Type":"ContainerStarted","Data":"a358757e51d1082b4ec6c33bd2e18d2538882b46f945b693893d15fbfd30282f"} Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.755748 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfpb" event={"ID":"ffb878ee-7cd7-4830-bb19-6b2a5e6bd1e0","Type":"ContainerStarted","Data":"6e8272e5958c82cf9a45a0a0c3c2866c78b423f8f8ddace664b5bb9d84e718e2"} Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.767685 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-h5tfm" event={"ID":"f62330e9-cdc5-469b-82ba-3364ce173686","Type":"ContainerStarted","Data":"a09a3090fa8105f004986f42b5732ef97306023a4bab7b6c8a5139c4b065d85a"} Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.770870 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" event={"ID":"5f37526d-ea7c-4607-920a-ca63aa96184c","Type":"ContainerStarted","Data":"63c099de631eb958ac858b63b909c7a4ea802abf54198d5076f1512239135162"} Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.784058 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-44xzn" event={"ID":"528125c4-7d83-4707-ae0d-da40498497d0","Type":"ContainerStarted","Data":"7cc29aa2ece61cc0a3abd432ceef15c9d9bf2125d31b85002d68e0526d23bc0c"} Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.790049 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-khvnh" event={"ID":"ad264986-2840-4a2a-bee9-9ea192cdbe86","Type":"ContainerStarted","Data":"509a944f1e8829ef2c2200054d21e814377d3da78a17f3253bb1b6ef47c6fe06"} Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.824416 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:46 crc kubenswrapper[4696]: E0321 08:29:46.824726 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:47.324700928 +0000 UTC m=+121.445581641 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.825040 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:46 crc kubenswrapper[4696]: E0321 08:29:46.825841 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:47.325805767 +0000 UTC m=+121.446686480 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.912609 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-bp44p" podStartSLOduration=80.911665661 podStartE2EDuration="1m20.911665661s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:46.901355192 +0000 UTC m=+121.022235905" watchObservedRunningTime="2026-03-21 08:29:46.911665661 +0000 UTC m=+121.032546374" Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.926300 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:46 crc kubenswrapper[4696]: E0321 08:29:46.928860 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:47.428839157 +0000 UTC m=+121.549719870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.943246 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-c9v5l" podStartSLOduration=80.943224521 podStartE2EDuration="1m20.943224521s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:46.941581409 +0000 UTC m=+121.062462132" watchObservedRunningTime="2026-03-21 08:29:46.943224521 +0000 UTC m=+121.064105234" Mar 21 08:29:46 crc kubenswrapper[4696]: I0321 08:29:46.985184 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xt5jj" podStartSLOduration=80.985160532 podStartE2EDuration="1m20.985160532s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:46.984371842 +0000 UTC m=+121.105252555" watchObservedRunningTime="2026-03-21 08:29:46.985160532 +0000 UTC m=+121.106041245" Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.031500 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:47 crc kubenswrapper[4696]: E0321 08:29:47.031882 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:47.531865387 +0000 UTC m=+121.652746110 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.050772 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" podStartSLOduration=81.050753999 podStartE2EDuration="1m21.050753999s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:47.009724002 +0000 UTC m=+121.130604705" watchObservedRunningTime="2026-03-21 08:29:47.050753999 +0000 UTC m=+121.171634702" Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.051456 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-9thjt" podStartSLOduration=81.051425806 podStartE2EDuration="1m21.051425806s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:47.04079462 +0000 UTC m=+121.161675353" watchObservedRunningTime="2026-03-21 08:29:47.051425806 +0000 UTC m=+121.172306519" Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.075813 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5thcd" podStartSLOduration=81.07578858 podStartE2EDuration="1m21.07578858s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:47.074509537 +0000 UTC m=+121.195390250" watchObservedRunningTime="2026-03-21 08:29:47.07578858 +0000 UTC m=+121.196669293" Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.114034 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r" podStartSLOduration=80.114017374 podStartE2EDuration="1m20.114017374s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:47.1126765 +0000 UTC m=+121.233557213" watchObservedRunningTime="2026-03-21 08:29:47.114017374 +0000 UTC m=+121.234898087" Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.135251 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:47 crc kubenswrapper[4696]: E0321 08:29:47.135951 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:47.635935155 +0000 UTC m=+121.756815878 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.247793 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:47 crc kubenswrapper[4696]: E0321 08:29:47.248685 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:47.748670287 +0000 UTC m=+121.869551000 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.280224 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-9thjt" Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.349353 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:47 crc kubenswrapper[4696]: E0321 08:29:47.349675 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:47.849659964 +0000 UTC m=+121.970540677 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.450706 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:47 crc kubenswrapper[4696]: E0321 08:29:47.451360 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:47.95133433 +0000 UTC m=+122.072215043 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.553394 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:47 crc kubenswrapper[4696]: E0321 08:29:47.553836 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:48.053798065 +0000 UTC m=+122.174678778 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:47 crc kubenswrapper[4696]: E0321 08:29:47.566508 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b9d4194_45b6_445b_b882_1982630eaecf.slice/crio-70857a43c6d20bd7a7757d96ef3bbb36bd10eef1c9854eb1ea71af3fa381359f\": RecentStats: unable to find data in memory cache]" Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.663237 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:47 crc kubenswrapper[4696]: E0321 08:29:47.664340 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:48.16432546 +0000 UTC m=+122.285206173 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.766759 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:47 crc kubenswrapper[4696]: E0321 08:29:47.767574 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:48.267553986 +0000 UTC m=+122.388434719 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.806292 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-4tb7m"] Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.842109 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-h5tfm" event={"ID":"f62330e9-cdc5-469b-82ba-3364ce173686","Type":"ContainerStarted","Data":"2e866313fc649545ffe87a19a3d0ba8533ca6bfe9b751fb1d70be9d8e7221ca6"} Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.847529 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-jdfkn"] Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.856086 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-xblff" event={"ID":"e24de32f-7778-4fac-815b-fa4d2a44dd2f","Type":"ContainerStarted","Data":"50f6142327323da65008b8ee8a750af4b8c5c012fe5d0b7626b2ccdce74fdace"} Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.858682 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wshd6"] Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.866455 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qc5tq" event={"ID":"dc140eb2-a579-45bc-9fb4-247c254533ea","Type":"ContainerStarted","Data":"5451baf096c7fc9499d053fb56295221b5522e208eb152d3dbf6d27636445549"} Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.869565 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:47 crc kubenswrapper[4696]: E0321 08:29:47.870015 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:48.370001191 +0000 UTC m=+122.490881904 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.890544 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-h5tfm" podStartSLOduration=80.890522675 podStartE2EDuration="1m20.890522675s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:47.88882455 +0000 UTC m=+122.009705273" watchObservedRunningTime="2026-03-21 08:29:47.890522675 +0000 UTC m=+122.011403388" Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.898398 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5mlz9"] Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.901888 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qm5pb"] Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.903980 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ntmrq"] Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.916563 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r656w"] Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.919944 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8hcm" event={"ID":"b999de84-1fbb-4184-a136-ce68641fc110","Type":"ContainerStarted","Data":"553a8f40806de0cc0d0f521d6a26a1a96de99d9d13cd2a7378fec289657e7b78"} Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.920021 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-xblff" podStartSLOduration=81.920003512 podStartE2EDuration="1m21.920003512s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:47.919217591 +0000 UTC m=+122.040098304" watchObservedRunningTime="2026-03-21 08:29:47.920003512 +0000 UTC m=+122.040884225" Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.929077 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-chtnc"] Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.941556 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48" event={"ID":"6ef60c27-2062-4bc8-8c37-e13db97d29a3","Type":"ContainerStarted","Data":"5a46b0e020a5fc8eeada2c072e338cc86a14724276bce439f8b313ba08e6bac6"} Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.943857 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qc5tq" podStartSLOduration=81.943836492 podStartE2EDuration="1m21.943836492s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:47.941283215 +0000 UTC m=+122.062163928" watchObservedRunningTime="2026-03-21 08:29:47.943836492 +0000 UTC m=+122.064717205" Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.948892 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl" event={"ID":"ed334ee3-b168-4ac1-94a7-45aa902b9b32","Type":"ContainerStarted","Data":"ff9a83e2d59baccfc0e4371b6105d201a9639eef7f3c2449b62de7346d53e349"} Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.954786 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bbpqw" event={"ID":"5b9d4194-45b6-445b-b882-1982630eaecf","Type":"ContainerStarted","Data":"70857a43c6d20bd7a7757d96ef3bbb36bd10eef1c9854eb1ea71af3fa381359f"} Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.956482 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9s5ck" event={"ID":"1df1b2d7-6ec6-4b8c-95d6-d08bdf698974","Type":"ContainerStarted","Data":"dc1608a81177ed471822f0fcf237ec8d1dfff0b8ca9cdfcf35d1e0e16410b027"} Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.964136 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-bp44p" event={"ID":"f2a60c2e-1edd-4c96-b263-e84dec7ed4a2","Type":"ContainerStarted","Data":"3ad7e19345b20b063ccfa907773831a59713774de7e9e7742d66abfc90cb60a3"} Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.970395 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:47 crc kubenswrapper[4696]: E0321 08:29:47.970666 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:48.470484105 +0000 UTC m=+122.591364828 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:47 crc kubenswrapper[4696]: I0321 08:29:47.971297 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:47 crc kubenswrapper[4696]: E0321 08:29:47.974404 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:48.474387827 +0000 UTC m=+122.595268540 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:48 crc kubenswrapper[4696]: I0321 08:29:48.000706 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" event={"ID":"973fa2b5-c6ce-498f-9b2b-311e8dd34057","Type":"ContainerStarted","Data":"4f55d748904168540bea4ad08042f6ff645af26914eb9737ac872aaa9d1627ba"} Mar 21 08:29:48 crc kubenswrapper[4696]: I0321 08:29:48.018668 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" event={"ID":"91299f09-692d-4c4e-a83c-22fcd785c1ef","Type":"ContainerStarted","Data":"a27af2d67b4195cc7486a96cc560e7605f93094a6ec70acce076e39b20a3b65e"} Mar 21 08:29:48 crc kubenswrapper[4696]: I0321 08:29:48.047930 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-w95sn"] Mar 21 08:29:48 crc kubenswrapper[4696]: W0321 08:29:48.075541 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a77c875_c8d2_40e5_aabc_4f7ffcd29693.slice/crio-25e372aa5e6f7cf88683fb89b75e2da63635c2a04f93c11355d2d421439200be WatchSource:0}: Error finding container 25e372aa5e6f7cf88683fb89b75e2da63635c2a04f93c11355d2d421439200be: Status 404 returned error can't find the container with id 25e372aa5e6f7cf88683fb89b75e2da63635c2a04f93c11355d2d421439200be Mar 21 08:29:48 crc kubenswrapper[4696]: I0321 08:29:48.077803 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:48 crc kubenswrapper[4696]: E0321 08:29:48.080117 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:48.580094026 +0000 UTC m=+122.700974739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:48 crc kubenswrapper[4696]: I0321 08:29:48.087085 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" podUID="0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c" containerName="route-controller-manager" containerID="cri-o://3e773b872d98df4836577cc15d9015665074913bd0488072a4a9e16c3145fe27" gracePeriod=30 Mar 21 08:29:48 crc kubenswrapper[4696]: I0321 08:29:48.087335 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-khvnh" event={"ID":"ad264986-2840-4a2a-bee9-9ea192cdbe86","Type":"ContainerStarted","Data":"d5b9ea79cc629a9f2cc46ee6c886df78998ffce814fae8c8603badd5179d9e57"} Mar 21 08:29:48 crc kubenswrapper[4696]: I0321 08:29:48.087511 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" podUID="14fd07c2-17e1-4adb-87c8-459f1dcb41c2" containerName="controller-manager" containerID="cri-o://65eadb3b7870c00e93eedf571b1bfa6163cf9a08c6fadf8c64efba6e41e33426" gracePeriod=30 Mar 21 08:29:48 crc kubenswrapper[4696]: I0321 08:29:48.107049 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:29:48 crc kubenswrapper[4696]: I0321 08:29:48.125161 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h5n9r" Mar 21 08:29:48 crc kubenswrapper[4696]: I0321 08:29:48.125232 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xt5jj" Mar 21 08:29:48 crc kubenswrapper[4696]: I0321 08:29:48.125524 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-khvnh" podStartSLOduration=82.125505778 podStartE2EDuration="1m22.125505778s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:48.125195239 +0000 UTC m=+122.246075952" watchObservedRunningTime="2026-03-21 08:29:48.125505778 +0000 UTC m=+122.246386511" Mar 21 08:29:48 crc kubenswrapper[4696]: I0321 08:29:48.139909 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:48 crc kubenswrapper[4696]: I0321 08:29:48.140282 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:48 crc kubenswrapper[4696]: I0321 08:29:48.186950 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:48 crc kubenswrapper[4696]: E0321 08:29:48.188591 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:48.688572919 +0000 UTC m=+122.809453722 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.839421 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:49 crc kubenswrapper[4696]: E0321 08:29:49.839895 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:50.839880321 +0000 UTC m=+124.960761034 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.884696 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.888150 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qs7dj" event={"ID":"14c95e94-4284-423b-b1cf-8df8ef663d65","Type":"ContainerStarted","Data":"6a9429707a312f9dd74c6420ed6175cc9fa56cf7a9edb04c03fcee41f0147f25"} Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.889500 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48" event={"ID":"6ef60c27-2062-4bc8-8c37-e13db97d29a3","Type":"ContainerStarted","Data":"26fd12ef9eb9bd83c4e74724cce711f57f7190be8e20833ce132ed584b0c3116"} Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.898611 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-sdgwl" event={"ID":"231fd5a3-3560-4a89-a5be-59880089a8fe","Type":"ContainerStarted","Data":"a4520caafcd82311e068e6d472d72decf59fddd01a14c011d6bf712ca8559269"} Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.910502 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r656w" event={"ID":"89b55c04-e4f6-4cbe-8e92-5cf8726c1486","Type":"ContainerStarted","Data":"75c282edbf7e6da04d63495ea658e46161bacd9a85ed333904c4d1e10eddfa7a"} Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.912740 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ntmrq" event={"ID":"45a093d5-91a7-44bf-8072-fc24898b79e8","Type":"ContainerStarted","Data":"433197a47c64c10020e8815a929aa67db3ffaee5ef014bb53f3d112d9319d2a0"} Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.922645 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qs7dj" podStartSLOduration=82.922601928 podStartE2EDuration="1m22.922601928s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:49.917885216 +0000 UTC m=+124.038765919" watchObservedRunningTime="2026-03-21 08:29:49.922601928 +0000 UTC m=+124.043482641" Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.935190 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wshd6" event={"ID":"9d4f19c5-4861-4ae6-99d0-d32e07de56e2","Type":"ContainerStarted","Data":"d8ac36deb761cdfa68e2fc636e0241993d912f0647299043874da0bf7097dee5"} Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.940497 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qsg67"] Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.941281 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.942000 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qsg67" Mar 21 08:29:49 crc kubenswrapper[4696]: E0321 08:29:49.943601 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:50.443552473 +0000 UTC m=+124.564433186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.947050 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v9rlg"] Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.948280 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5mlz9" event={"ID":"6769f68b-1f58-48e4-8076-14fcdb9c8036","Type":"ContainerStarted","Data":"849464c2127d9c902aa792bc6f5e3ff32e1ccae991b9b6c84808e867248ceaf0"} Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.948398 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v9rlg" Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.954717 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7wvkn"] Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.965292 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.965937 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.966758 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7wvkn" Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.970189 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=0.970163965 podStartE2EDuration="970.163965ms" podCreationTimestamp="2026-03-21 08:29:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:49.966205612 +0000 UTC m=+124.087086335" watchObservedRunningTime="2026-03-21 08:29:49.970163965 +0000 UTC m=+124.091044698" Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.972759 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qsg67"] Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.977828 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-jdfkn" event={"ID":"f3f5d127-addf-4f3a-a9f0-fcb502db57c0","Type":"ContainerStarted","Data":"0b9f5a2ce947bef2f447d2c6ce2b6bb5c0c9b85075425454a63b5192c508b844"} Mar 21 08:29:49 crc kubenswrapper[4696]: I0321 08:29:49.980632 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7tpgm"] Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.007748 4696 generic.go:334] "Generic (PLEG): container finished" podID="0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c" containerID="3e773b872d98df4836577cc15d9015665074913bd0488072a4a9e16c3145fe27" exitCode=0 Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.009442 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-chtnc" event={"ID":"367f9b3a-40fe-414e-bc06-949205afe2b8","Type":"ContainerStarted","Data":"05b0fccbd206d24063ed84c5543ab0f0519a846e51c5b88d61fa4733885552d4"} Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.009469 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" event={"ID":"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c","Type":"ContainerDied","Data":"3e773b872d98df4836577cc15d9015665074913bd0488072a4a9e16c3145fe27"} Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.009586 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7tpgm" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.017223 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-w95sn" event={"ID":"9bbfb96a-607b-47df-83df-7c75dd94ad71","Type":"ContainerStarted","Data":"f71c15398e8769c9a79f27388a6e5d080b64352dd8ba30f7158fc75953b27a36"} Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.020665 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7wvkn"] Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.022418 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qm5pb" event={"ID":"8a77c875-c8d2-40e5-aabc-4f7ffcd29693","Type":"ContainerStarted","Data":"25e372aa5e6f7cf88683fb89b75e2da63635c2a04f93c11355d2d421439200be"} Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.031397 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7tpgm"] Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.039245 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-4tb7m" event={"ID":"6ec15f5e-2f88-4f9c-b581-d6ec874f6337","Type":"ContainerStarted","Data":"525ca67c2d5f382eec432ef9626dee35cdff242b453c7f3614042380c3212e92"} Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.042942 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85a5000b-3b68-4808-bcfc-93bcbde10dde-utilities\") pod \"community-operators-v9rlg\" (UID: \"85a5000b-3b68-4808-bcfc-93bcbde10dde\") " pod="openshift-marketplace/community-operators-v9rlg" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.043053 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zxpn\" (UniqueName: \"kubernetes.io/projected/85a5000b-3b68-4808-bcfc-93bcbde10dde-kube-api-access-4zxpn\") pod \"community-operators-v9rlg\" (UID: \"85a5000b-3b68-4808-bcfc-93bcbde10dde\") " pod="openshift-marketplace/community-operators-v9rlg" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.043127 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85a5000b-3b68-4808-bcfc-93bcbde10dde-catalog-content\") pod \"community-operators-v9rlg\" (UID: \"85a5000b-3b68-4808-bcfc-93bcbde10dde\") " pod="openshift-marketplace/community-operators-v9rlg" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.043151 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6aabaca2-7def-4550-bae1-9337b7c1002b-catalog-content\") pod \"certified-operators-qsg67\" (UID: \"6aabaca2-7def-4550-bae1-9337b7c1002b\") " pod="openshift-marketplace/certified-operators-qsg67" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.043230 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b88fc0a-dc29-4aa2-b625-c24162c0dc0b-utilities\") pod \"community-operators-7wvkn\" (UID: \"0b88fc0a-dc29-4aa2-b625-c24162c0dc0b\") " pod="openshift-marketplace/community-operators-7wvkn" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.043305 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5149cc5e-d877-4d52-8db6-7df6ae060ab0-utilities\") pod \"certified-operators-7tpgm\" (UID: \"5149cc5e-d877-4d52-8db6-7df6ae060ab0\") " pod="openshift-marketplace/certified-operators-7tpgm" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.043330 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5149cc5e-d877-4d52-8db6-7df6ae060ab0-catalog-content\") pod \"certified-operators-7tpgm\" (UID: \"5149cc5e-d877-4d52-8db6-7df6ae060ab0\") " pod="openshift-marketplace/certified-operators-7tpgm" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.043380 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b88fc0a-dc29-4aa2-b625-c24162c0dc0b-catalog-content\") pod \"community-operators-7wvkn\" (UID: \"0b88fc0a-dc29-4aa2-b625-c24162c0dc0b\") " pod="openshift-marketplace/community-operators-7wvkn" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.043400 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6aabaca2-7def-4550-bae1-9337b7c1002b-utilities\") pod \"certified-operators-qsg67\" (UID: \"6aabaca2-7def-4550-bae1-9337b7c1002b\") " pod="openshift-marketplace/certified-operators-qsg67" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.043520 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.043563 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gs9s\" (UniqueName: \"kubernetes.io/projected/5149cc5e-d877-4d52-8db6-7df6ae060ab0-kube-api-access-7gs9s\") pod \"certified-operators-7tpgm\" (UID: \"5149cc5e-d877-4d52-8db6-7df6ae060ab0\") " pod="openshift-marketplace/certified-operators-7tpgm" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.043582 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jth8k\" (UniqueName: \"kubernetes.io/projected/0b88fc0a-dc29-4aa2-b625-c24162c0dc0b-kube-api-access-jth8k\") pod \"community-operators-7wvkn\" (UID: \"0b88fc0a-dc29-4aa2-b625-c24162c0dc0b\") " pod="openshift-marketplace/community-operators-7wvkn" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.043621 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gswrs\" (UniqueName: \"kubernetes.io/projected/6aabaca2-7def-4550-bae1-9337b7c1002b-kube-api-access-gswrs\") pod \"certified-operators-qsg67\" (UID: \"6aabaca2-7def-4550-bae1-9337b7c1002b\") " pod="openshift-marketplace/certified-operators-qsg67" Mar 21 08:29:50 crc kubenswrapper[4696]: E0321 08:29:50.044007 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:50.543994976 +0000 UTC m=+124.664875679 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.045405 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd5f5" event={"ID":"b4928d31-a7c1-47f0-a726-712de834e034","Type":"ContainerStarted","Data":"0a7e016f0c30e683869d33d7e059cbb78e7c3cde5bbf747786543d52b346eefd"} Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.053136 4696 generic.go:334] "Generic (PLEG): container finished" podID="14fd07c2-17e1-4adb-87c8-459f1dcb41c2" containerID="65eadb3b7870c00e93eedf571b1bfa6163cf9a08c6fadf8c64efba6e41e33426" exitCode=0 Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.055127 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" event={"ID":"14fd07c2-17e1-4adb-87c8-459f1dcb41c2","Type":"ContainerDied","Data":"65eadb3b7870c00e93eedf571b1bfa6163cf9a08c6fadf8c64efba6e41e33426"} Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.055584 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v9rlg"] Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.096257 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-sdgwl" podStartSLOduration=83.096231915 podStartE2EDuration="1m23.096231915s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:50.040792952 +0000 UTC m=+124.161673665" watchObservedRunningTime="2026-03-21 08:29:50.096231915 +0000 UTC m=+124.217112628" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.098382 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48" podStartSLOduration=84.098371011 podStartE2EDuration="1m24.098371011s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:50.063384711 +0000 UTC m=+124.184265424" watchObservedRunningTime="2026-03-21 08:29:50.098371011 +0000 UTC m=+124.219251754" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.145292 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.145697 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85a5000b-3b68-4808-bcfc-93bcbde10dde-utilities\") pod \"community-operators-v9rlg\" (UID: \"85a5000b-3b68-4808-bcfc-93bcbde10dde\") " pod="openshift-marketplace/community-operators-v9rlg" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.145788 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zxpn\" (UniqueName: \"kubernetes.io/projected/85a5000b-3b68-4808-bcfc-93bcbde10dde-kube-api-access-4zxpn\") pod \"community-operators-v9rlg\" (UID: \"85a5000b-3b68-4808-bcfc-93bcbde10dde\") " pod="openshift-marketplace/community-operators-v9rlg" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.145812 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85a5000b-3b68-4808-bcfc-93bcbde10dde-catalog-content\") pod \"community-operators-v9rlg\" (UID: \"85a5000b-3b68-4808-bcfc-93bcbde10dde\") " pod="openshift-marketplace/community-operators-v9rlg" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.145854 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b88fc0a-dc29-4aa2-b625-c24162c0dc0b-utilities\") pod \"community-operators-7wvkn\" (UID: \"0b88fc0a-dc29-4aa2-b625-c24162c0dc0b\") " pod="openshift-marketplace/community-operators-7wvkn" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.145877 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6aabaca2-7def-4550-bae1-9337b7c1002b-catalog-content\") pod \"certified-operators-qsg67\" (UID: \"6aabaca2-7def-4550-bae1-9337b7c1002b\") " pod="openshift-marketplace/certified-operators-qsg67" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.146030 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5149cc5e-d877-4d52-8db6-7df6ae060ab0-utilities\") pod \"certified-operators-7tpgm\" (UID: \"5149cc5e-d877-4d52-8db6-7df6ae060ab0\") " pod="openshift-marketplace/certified-operators-7tpgm" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.146086 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5149cc5e-d877-4d52-8db6-7df6ae060ab0-catalog-content\") pod \"certified-operators-7tpgm\" (UID: \"5149cc5e-d877-4d52-8db6-7df6ae060ab0\") " pod="openshift-marketplace/certified-operators-7tpgm" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.146128 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b88fc0a-dc29-4aa2-b625-c24162c0dc0b-catalog-content\") pod \"community-operators-7wvkn\" (UID: \"0b88fc0a-dc29-4aa2-b625-c24162c0dc0b\") " pod="openshift-marketplace/community-operators-7wvkn" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.146149 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6aabaca2-7def-4550-bae1-9337b7c1002b-utilities\") pod \"certified-operators-qsg67\" (UID: \"6aabaca2-7def-4550-bae1-9337b7c1002b\") " pod="openshift-marketplace/certified-operators-qsg67" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.146316 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gs9s\" (UniqueName: \"kubernetes.io/projected/5149cc5e-d877-4d52-8db6-7df6ae060ab0-kube-api-access-7gs9s\") pod \"certified-operators-7tpgm\" (UID: \"5149cc5e-d877-4d52-8db6-7df6ae060ab0\") " pod="openshift-marketplace/certified-operators-7tpgm" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.146347 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jth8k\" (UniqueName: \"kubernetes.io/projected/0b88fc0a-dc29-4aa2-b625-c24162c0dc0b-kube-api-access-jth8k\") pod \"community-operators-7wvkn\" (UID: \"0b88fc0a-dc29-4aa2-b625-c24162c0dc0b\") " pod="openshift-marketplace/community-operators-7wvkn" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.146450 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gswrs\" (UniqueName: \"kubernetes.io/projected/6aabaca2-7def-4550-bae1-9337b7c1002b-kube-api-access-gswrs\") pod \"certified-operators-qsg67\" (UID: \"6aabaca2-7def-4550-bae1-9337b7c1002b\") " pod="openshift-marketplace/certified-operators-qsg67" Mar 21 08:29:50 crc kubenswrapper[4696]: E0321 08:29:50.147863 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:50.647847528 +0000 UTC m=+124.768728241 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.149873 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b88fc0a-dc29-4aa2-b625-c24162c0dc0b-catalog-content\") pod \"community-operators-7wvkn\" (UID: \"0b88fc0a-dc29-4aa2-b625-c24162c0dc0b\") " pod="openshift-marketplace/community-operators-7wvkn" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.150870 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85a5000b-3b68-4808-bcfc-93bcbde10dde-catalog-content\") pod \"community-operators-v9rlg\" (UID: \"85a5000b-3b68-4808-bcfc-93bcbde10dde\") " pod="openshift-marketplace/community-operators-v9rlg" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.151182 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b88fc0a-dc29-4aa2-b625-c24162c0dc0b-utilities\") pod \"community-operators-7wvkn\" (UID: \"0b88fc0a-dc29-4aa2-b625-c24162c0dc0b\") " pod="openshift-marketplace/community-operators-7wvkn" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.153399 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5149cc5e-d877-4d52-8db6-7df6ae060ab0-catalog-content\") pod \"certified-operators-7tpgm\" (UID: \"5149cc5e-d877-4d52-8db6-7df6ae060ab0\") " pod="openshift-marketplace/certified-operators-7tpgm" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.153480 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85a5000b-3b68-4808-bcfc-93bcbde10dde-utilities\") pod \"community-operators-v9rlg\" (UID: \"85a5000b-3b68-4808-bcfc-93bcbde10dde\") " pod="openshift-marketplace/community-operators-v9rlg" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.159213 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5149cc5e-d877-4d52-8db6-7df6ae060ab0-utilities\") pod \"certified-operators-7tpgm\" (UID: \"5149cc5e-d877-4d52-8db6-7df6ae060ab0\") " pod="openshift-marketplace/certified-operators-7tpgm" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.164181 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6aabaca2-7def-4550-bae1-9337b7c1002b-catalog-content\") pod \"certified-operators-qsg67\" (UID: \"6aabaca2-7def-4550-bae1-9337b7c1002b\") " pod="openshift-marketplace/certified-operators-qsg67" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.165192 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6aabaca2-7def-4550-bae1-9337b7c1002b-utilities\") pod \"certified-operators-qsg67\" (UID: \"6aabaca2-7def-4550-bae1-9337b7c1002b\") " pod="openshift-marketplace/certified-operators-qsg67" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.182738 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gswrs\" (UniqueName: \"kubernetes.io/projected/6aabaca2-7def-4550-bae1-9337b7c1002b-kube-api-access-gswrs\") pod \"certified-operators-qsg67\" (UID: \"6aabaca2-7def-4550-bae1-9337b7c1002b\") " pod="openshift-marketplace/certified-operators-qsg67" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.183025 4696 patch_prober.go:28] interesting pod/apiserver-76f77b778f-xblff container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 21 08:29:50 crc kubenswrapper[4696]: [+]log ok Mar 21 08:29:50 crc kubenswrapper[4696]: [+]etcd ok Mar 21 08:29:50 crc kubenswrapper[4696]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 21 08:29:50 crc kubenswrapper[4696]: [+]poststarthook/generic-apiserver-start-informers ok Mar 21 08:29:50 crc kubenswrapper[4696]: [+]poststarthook/max-in-flight-filter ok Mar 21 08:29:50 crc kubenswrapper[4696]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 21 08:29:50 crc kubenswrapper[4696]: [+]poststarthook/image.openshift.io-apiserver-caches ok Mar 21 08:29:50 crc kubenswrapper[4696]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Mar 21 08:29:50 crc kubenswrapper[4696]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Mar 21 08:29:50 crc kubenswrapper[4696]: [+]poststarthook/project.openshift.io-projectcache ok Mar 21 08:29:50 crc kubenswrapper[4696]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Mar 21 08:29:50 crc kubenswrapper[4696]: [+]poststarthook/openshift.io-startinformers ok Mar 21 08:29:50 crc kubenswrapper[4696]: [+]poststarthook/openshift.io-restmapperupdater ok Mar 21 08:29:50 crc kubenswrapper[4696]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Mar 21 08:29:50 crc kubenswrapper[4696]: livez check failed Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.183078 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-xblff" podUID="e24de32f-7778-4fac-815b-fa4d2a44dd2f" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.188696 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zxpn\" (UniqueName: \"kubernetes.io/projected/85a5000b-3b68-4808-bcfc-93bcbde10dde-kube-api-access-4zxpn\") pod \"community-operators-v9rlg\" (UID: \"85a5000b-3b68-4808-bcfc-93bcbde10dde\") " pod="openshift-marketplace/community-operators-v9rlg" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.192498 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jth8k\" (UniqueName: \"kubernetes.io/projected/0b88fc0a-dc29-4aa2-b625-c24162c0dc0b-kube-api-access-jth8k\") pod \"community-operators-7wvkn\" (UID: \"0b88fc0a-dc29-4aa2-b625-c24162c0dc0b\") " pod="openshift-marketplace/community-operators-7wvkn" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.193751 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gs9s\" (UniqueName: \"kubernetes.io/projected/5149cc5e-d877-4d52-8db6-7df6ae060ab0-kube-api-access-7gs9s\") pod \"certified-operators-7tpgm\" (UID: \"5149cc5e-d877-4d52-8db6-7df6ae060ab0\") " pod="openshift-marketplace/certified-operators-7tpgm" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.247475 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:50 crc kubenswrapper[4696]: E0321 08:29:50.247829 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:50.747800608 +0000 UTC m=+124.868681321 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.248597 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd5f5" podStartSLOduration=83.248576898 podStartE2EDuration="1m23.248576898s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:50.247650474 +0000 UTC m=+124.368531187" watchObservedRunningTime="2026-03-21 08:29:50.248576898 +0000 UTC m=+124.369457611" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.353009 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:50 crc kubenswrapper[4696]: E0321 08:29:50.353435 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:50.853414436 +0000 UTC m=+124.974295149 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.456522 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:50 crc kubenswrapper[4696]: E0321 08:29:50.456898 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:50.956882757 +0000 UTC m=+125.077763470 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.469175 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qsg67" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.532389 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v9rlg" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.559100 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:50 crc kubenswrapper[4696]: E0321 08:29:50.560728 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:51.060648137 +0000 UTC m=+125.181528850 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.573678 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7wvkn" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.595798 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7tpgm" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.603278 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.614593 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.653895 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj"] Mar 21 08:29:50 crc kubenswrapper[4696]: E0321 08:29:50.654153 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14fd07c2-17e1-4adb-87c8-459f1dcb41c2" containerName="controller-manager" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.654168 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="14fd07c2-17e1-4adb-87c8-459f1dcb41c2" containerName="controller-manager" Mar 21 08:29:50 crc kubenswrapper[4696]: E0321 08:29:50.654184 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c" containerName="route-controller-manager" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.654192 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c" containerName="route-controller-manager" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.654300 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="14fd07c2-17e1-4adb-87c8-459f1dcb41c2" containerName="controller-manager" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.654311 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c" containerName="route-controller-manager" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.654659 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.666902 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-serving-cert\") pod \"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c\" (UID: \"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c\") " Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.666933 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-config\") pod \"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c\" (UID: \"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c\") " Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.666985 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvhrn\" (UniqueName: \"kubernetes.io/projected/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-kube-api-access-qvhrn\") pod \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\" (UID: \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\") " Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.667005 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-proxy-ca-bundles\") pod \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\" (UID: \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\") " Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.667070 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-serving-cert\") pod \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\" (UID: \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\") " Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.667093 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-client-ca\") pod \"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c\" (UID: \"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c\") " Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.667114 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-client-ca\") pod \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\" (UID: \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\") " Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.667138 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-config\") pod \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\" (UID: \"14fd07c2-17e1-4adb-87c8-459f1dcb41c2\") " Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.667260 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lc6gp\" (UniqueName: \"kubernetes.io/projected/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-kube-api-access-lc6gp\") pod \"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c\" (UID: \"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c\") " Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.667522 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1690a33-2e30-4c37-ba60-5def18cc3867-config\") pod \"route-controller-manager-847bc8884c-559kj\" (UID: \"e1690a33-2e30-4c37-ba60-5def18cc3867\") " pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.667555 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e1690a33-2e30-4c37-ba60-5def18cc3867-client-ca\") pod \"route-controller-manager-847bc8884c-559kj\" (UID: \"e1690a33-2e30-4c37-ba60-5def18cc3867\") " pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.667575 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1690a33-2e30-4c37-ba60-5def18cc3867-serving-cert\") pod \"route-controller-manager-847bc8884c-559kj\" (UID: \"e1690a33-2e30-4c37-ba60-5def18cc3867\") " pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.667601 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh7kh\" (UniqueName: \"kubernetes.io/projected/e1690a33-2e30-4c37-ba60-5def18cc3867-kube-api-access-wh7kh\") pod \"route-controller-manager-847bc8884c-559kj\" (UID: \"e1690a33-2e30-4c37-ba60-5def18cc3867\") " pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.667637 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.668080 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-config" (OuterVolumeSpecName: "config") pod "0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c" (UID: "0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:50 crc kubenswrapper[4696]: E0321 08:29:50.669439 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:51.169421456 +0000 UTC m=+125.290302169 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.670656 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-client-ca" (OuterVolumeSpecName: "client-ca") pod "0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c" (UID: "0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.680490 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-config" (OuterVolumeSpecName: "config") pod "14fd07c2-17e1-4adb-87c8-459f1dcb41c2" (UID: "14fd07c2-17e1-4adb-87c8-459f1dcb41c2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.681142 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "14fd07c2-17e1-4adb-87c8-459f1dcb41c2" (UID: "14fd07c2-17e1-4adb-87c8-459f1dcb41c2"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.681723 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-kube-api-access-qvhrn" (OuterVolumeSpecName: "kube-api-access-qvhrn") pod "14fd07c2-17e1-4adb-87c8-459f1dcb41c2" (UID: "14fd07c2-17e1-4adb-87c8-459f1dcb41c2"). InnerVolumeSpecName "kube-api-access-qvhrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.681781 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-client-ca" (OuterVolumeSpecName: "client-ca") pod "14fd07c2-17e1-4adb-87c8-459f1dcb41c2" (UID: "14fd07c2-17e1-4adb-87c8-459f1dcb41c2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.692932 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj"] Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.702622 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-sdgwl" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.724184 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-kube-api-access-lc6gp" (OuterVolumeSpecName: "kube-api-access-lc6gp") pod "0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c" (UID: "0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c"). InnerVolumeSpecName "kube-api-access-lc6gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.724370 4696 patch_prober.go:28] interesting pod/router-default-5444994796-sdgwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 08:29:50 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Mar 21 08:29:50 crc kubenswrapper[4696]: [+]process-running ok Mar 21 08:29:50 crc kubenswrapper[4696]: healthz check failed Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.724430 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sdgwl" podUID="231fd5a3-3560-4a89-a5be-59880089a8fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.733389 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "14fd07c2-17e1-4adb-87c8-459f1dcb41c2" (UID: "14fd07c2-17e1-4adb-87c8-459f1dcb41c2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.739805 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c" (UID: "0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.770552 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.770755 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1690a33-2e30-4c37-ba60-5def18cc3867-config\") pod \"route-controller-manager-847bc8884c-559kj\" (UID: \"e1690a33-2e30-4c37-ba60-5def18cc3867\") " pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.770785 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e1690a33-2e30-4c37-ba60-5def18cc3867-client-ca\") pod \"route-controller-manager-847bc8884c-559kj\" (UID: \"e1690a33-2e30-4c37-ba60-5def18cc3867\") " pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.770830 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1690a33-2e30-4c37-ba60-5def18cc3867-serving-cert\") pod \"route-controller-manager-847bc8884c-559kj\" (UID: \"e1690a33-2e30-4c37-ba60-5def18cc3867\") " pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.770858 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh7kh\" (UniqueName: \"kubernetes.io/projected/e1690a33-2e30-4c37-ba60-5def18cc3867-kube-api-access-wh7kh\") pod \"route-controller-manager-847bc8884c-559kj\" (UID: \"e1690a33-2e30-4c37-ba60-5def18cc3867\") " pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.770951 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvhrn\" (UniqueName: \"kubernetes.io/projected/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-kube-api-access-qvhrn\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.770961 4696 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.770970 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.770978 4696 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.770986 4696 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.770994 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14fd07c2-17e1-4adb-87c8-459f1dcb41c2-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.771003 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lc6gp\" (UniqueName: \"kubernetes.io/projected/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-kube-api-access-lc6gp\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.771011 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.771021 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:50 crc kubenswrapper[4696]: E0321 08:29:50.771364 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:51.271348177 +0000 UTC m=+125.392228890 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.772347 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1690a33-2e30-4c37-ba60-5def18cc3867-config\") pod \"route-controller-manager-847bc8884c-559kj\" (UID: \"e1690a33-2e30-4c37-ba60-5def18cc3867\") " pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.827008 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e1690a33-2e30-4c37-ba60-5def18cc3867-client-ca\") pod \"route-controller-manager-847bc8884c-559kj\" (UID: \"e1690a33-2e30-4c37-ba60-5def18cc3867\") " pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.830683 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1690a33-2e30-4c37-ba60-5def18cc3867-serving-cert\") pod \"route-controller-manager-847bc8884c-559kj\" (UID: \"e1690a33-2e30-4c37-ba60-5def18cc3867\") " pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.833396 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh7kh\" (UniqueName: \"kubernetes.io/projected/e1690a33-2e30-4c37-ba60-5def18cc3867-kube-api-access-wh7kh\") pod \"route-controller-manager-847bc8884c-559kj\" (UID: \"e1690a33-2e30-4c37-ba60-5def18cc3867\") " pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.877981 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:50 crc kubenswrapper[4696]: E0321 08:29:50.878351 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:51.378338622 +0000 UTC m=+125.499219335 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:50 crc kubenswrapper[4696]: I0321 08:29:50.983339 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:50 crc kubenswrapper[4696]: E0321 08:29:50.983888 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:51.483868346 +0000 UTC m=+125.604749059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.025766 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.085122 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:51 crc kubenswrapper[4696]: E0321 08:29:51.085520 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:51.585505651 +0000 UTC m=+125.706386364 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.138957 4696 ???:1] "http: TLS handshake error from 192.168.126.11:37440: no serving certificate available for the kubelet" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.177765 4696 generic.go:334] "Generic (PLEG): container finished" podID="91299f09-692d-4c4e-a83c-22fcd785c1ef" containerID="a1c4b02b4f97ad050a13b2725f4ec9f719a95e3172126bd4b1dd3dffb6d07a96" exitCode=0 Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.177929 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" event={"ID":"91299f09-692d-4c4e-a83c-22fcd785c1ef","Type":"ContainerDied","Data":"a1c4b02b4f97ad050a13b2725f4ec9f719a95e3172126bd4b1dd3dffb6d07a96"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.186282 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:51 crc kubenswrapper[4696]: E0321 08:29:51.186687 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:51.686668702 +0000 UTC m=+125.807549415 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.189777 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r656w" event={"ID":"89b55c04-e4f6-4cbe-8e92-5cf8726c1486","Type":"ContainerStarted","Data":"a94c516a5d9ff316e81df970e17ea1af63e59aa8a013b5603f36714bb7a9e86b"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.212381 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-jdfkn" event={"ID":"f3f5d127-addf-4f3a-a9f0-fcb502db57c0","Type":"ContainerStarted","Data":"c2a87246d20e3e7d9a25067a52b05af7fb2030da75b2ef210eb676b904a4a535"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.214122 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-jdfkn" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.228171 4696 ???:1] "http: TLS handshake error from 192.168.126.11:37450: no serving certificate available for the kubelet" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.256384 4696 patch_prober.go:28] interesting pod/downloads-7954f5f757-jdfkn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.256571 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-jdfkn" podUID="f3f5d127-addf-4f3a-a9f0-fcb502db57c0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.258444 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-65h6c" event={"ID":"2390b3ae-cb4b-4e89-854e-af5c0cf04e1d","Type":"ContainerStarted","Data":"10f57b95406edde8f339c16b0aae1381c2fd917f87901a08b27c94eb88f0c4fb"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.286613 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qm5pb" event={"ID":"8a77c875-c8d2-40e5-aabc-4f7ffcd29693","Type":"ContainerStarted","Data":"2a2fae9b6188a78b8cea75bbd4ac81b66c73e5be33f26dac05894b94a9183b65"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.287457 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:51 crc kubenswrapper[4696]: E0321 08:29:51.288638 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:51.788608304 +0000 UTC m=+125.909489017 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.289861 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-jdfkn" podStartSLOduration=85.289850357 podStartE2EDuration="1m25.289850357s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:51.271719315 +0000 UTC m=+125.392600028" watchObservedRunningTime="2026-03-21 08:29:51.289850357 +0000 UTC m=+125.410731070" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.319848 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qsg67"] Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.349163 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r656w" podStartSLOduration=84.349148789 podStartE2EDuration="1m24.349148789s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:51.316022987 +0000 UTC m=+125.436903700" watchObservedRunningTime="2026-03-21 08:29:51.349148789 +0000 UTC m=+125.470029492" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.356992 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" event={"ID":"934c62b8-c7a7-4018-b5f1-7954b2baca00","Type":"ContainerStarted","Data":"67d495ca6b81a3863c029a6fc0c294ab4d75c4562b0b1bb8e7c043f84882ba79"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.358195 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.358198 4696 ???:1] "http: TLS handshake error from 192.168.126.11:37456: no serving certificate available for the kubelet" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.390539 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:51 crc kubenswrapper[4696]: E0321 08:29:51.391723 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:51.891708966 +0000 UTC m=+126.012589679 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:51 crc kubenswrapper[4696]: W0321 08:29:51.400703 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6aabaca2_7def_4550_bae1_9337b7c1002b.slice/crio-ab7fb5221a42f276201e0f172904818436768bf0492bc00ee90713be326befaf WatchSource:0}: Error finding container ab7fb5221a42f276201e0f172904818436768bf0492bc00ee90713be326befaf: Status 404 returned error can't find the container with id ab7fb5221a42f276201e0f172904818436768bf0492bc00ee90713be326befaf Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.408152 4696 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-h7fjx container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" start-of-body= Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.408390 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" podUID="934c62b8-c7a7-4018-b5f1-7954b2baca00" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.431205 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-65h6c" podStartSLOduration=84.431191393 podStartE2EDuration="1m24.431191393s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:51.377304381 +0000 UTC m=+125.498185094" watchObservedRunningTime="2026-03-21 08:29:51.431191393 +0000 UTC m=+125.552072106" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.439877 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" podStartSLOduration=84.439850469 podStartE2EDuration="1m24.439850469s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:51.429797756 +0000 UTC m=+125.550678479" watchObservedRunningTime="2026-03-21 08:29:51.439850469 +0000 UTC m=+125.560731182" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.465266 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ntmrq" event={"ID":"45a093d5-91a7-44bf-8072-fc24898b79e8","Type":"ContainerStarted","Data":"c5b7952ba13b4ec0833bb205cb85955d62de4dbc203fec9e170715ff9cae062d"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.496167 4696 ???:1] "http: TLS handshake error from 192.168.126.11:37472: no serving certificate available for the kubelet" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.497006 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:51 crc kubenswrapper[4696]: E0321 08:29:51.497280 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:51.997267102 +0000 UTC m=+126.118147815 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.507654 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-6m4ch" event={"ID":"e26909f3-14e6-4f5a-9ea9-5b44ac3aa56a","Type":"ContainerStarted","Data":"e05c437d4f9450ff479cd68d1e47e489c7d4040d8c5ee5cca4939f87c623e1f5"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.543068 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-44xzn" event={"ID":"528125c4-7d83-4707-ae0d-da40498497d0","Type":"ContainerStarted","Data":"433e30b0966d41157625359a0c56bc1b26f65d4f6d22efe252e2dec07ec26b1c"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.545519 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-4tb7m" event={"ID":"6ec15f5e-2f88-4f9c-b581-d6ec874f6337","Type":"ContainerStarted","Data":"8c6249c374210ea52a02d5ae21efb2758280e76b7e45a3d6ca41768c34600af6"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.562711 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-6m4ch" podStartSLOduration=85.562692244 podStartE2EDuration="1m25.562692244s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:51.552666413 +0000 UTC m=+125.673547126" watchObservedRunningTime="2026-03-21 08:29:51.562692244 +0000 UTC m=+125.683572957" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.579099 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" event={"ID":"0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c","Type":"ContainerDied","Data":"5846194320be0a87de6399288a8b8b403658935759d629e5c55cbad3ebadc9de"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.579177 4696 scope.go:117] "RemoveContainer" containerID="3e773b872d98df4836577cc15d9015665074913bd0488072a4a9e16c3145fe27" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.579364 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.585077 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-44xzn" podStartSLOduration=10.585057895 podStartE2EDuration="10.585057895s" podCreationTimestamp="2026-03-21 08:29:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:51.579188613 +0000 UTC m=+125.700069316" watchObservedRunningTime="2026-03-21 08:29:51.585057895 +0000 UTC m=+125.705938608" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.585908 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v9rlg"] Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.598306 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:51 crc kubenswrapper[4696]: E0321 08:29:51.599928 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:52.099910032 +0000 UTC m=+126.220790745 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.600529 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7wvkn"] Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.612606 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-4tb7m" podStartSLOduration=10.612590052 podStartE2EDuration="10.612590052s" podCreationTimestamp="2026-03-21 08:29:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:51.610485998 +0000 UTC m=+125.731366711" watchObservedRunningTime="2026-03-21 08:29:51.612590052 +0000 UTC m=+125.733470765" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.617165 4696 ???:1] "http: TLS handshake error from 192.168.126.11:37474: no serving certificate available for the kubelet" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.622678 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.623329 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ptg5t" event={"ID":"14fd07c2-17e1-4adb-87c8-459f1dcb41c2","Type":"ContainerDied","Data":"1b330f2eb9e11f0e261c92ff501074e176cbf65d9bb598aa233d095589c70fe1"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.631686 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5mlz9" event={"ID":"6769f68b-1f58-48e4-8076-14fcdb9c8036","Type":"ContainerStarted","Data":"b8c3af66c2884843ff8414de9e4982eb65434c9d1cdf4c4fc13f9a26f5518559"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.657277 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-59mfl" event={"ID":"bf190816-3a09-4918-ab57-d31e55919f3e","Type":"ContainerStarted","Data":"915ca0d3130d7e47189d2ccd0d8567000068fb26a46c36b6a181857fa99ca231"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.669265 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6"] Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.676012 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xnns6"] Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.680907 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7tpgm"] Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.706701 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:51 crc kubenswrapper[4696]: E0321 08:29:51.707032 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:52.207019488 +0000 UTC m=+126.327900201 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.720567 4696 patch_prober.go:28] interesting pod/router-default-5444994796-sdgwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 08:29:51 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Mar 21 08:29:51 crc kubenswrapper[4696]: [+]process-running ok Mar 21 08:29:51 crc kubenswrapper[4696]: healthz check failed Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.720628 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sdgwl" podUID="231fd5a3-3560-4a89-a5be-59880089a8fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.721939 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5mlz9" podStartSLOduration=84.721926066 podStartE2EDuration="1m24.721926066s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:51.706323791 +0000 UTC m=+125.827204504" watchObservedRunningTime="2026-03-21 08:29:51.721926066 +0000 UTC m=+125.842806779" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.736487 4696 ???:1] "http: TLS handshake error from 192.168.126.11:37482: no serving certificate available for the kubelet" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.740241 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jnw8m"] Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.743271 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jnw8m" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.751296 4696 scope.go:117] "RemoveContainer" containerID="65eadb3b7870c00e93eedf571b1bfa6163cf9a08c6fadf8c64efba6e41e33426" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.755543 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.763278 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jnw8m"] Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.764560 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-59mfl" podStartSLOduration=84.764548936 podStartE2EDuration="1m24.764548936s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:51.745265444 +0000 UTC m=+125.866146157" watchObservedRunningTime="2026-03-21 08:29:51.764548936 +0000 UTC m=+125.885429659" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.765227 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl" event={"ID":"ed334ee3-b168-4ac1-94a7-45aa902b9b32","Type":"ContainerStarted","Data":"f8ddbe609a111b5fbc5539aff35d328c138f995885689db0c993886de7860923"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.765283 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl" event={"ID":"ed334ee3-b168-4ac1-94a7-45aa902b9b32","Type":"ContainerStarted","Data":"ef0fff7069f8eb7617a9676469520212af2b81448f5d225ee4fdd95a078c32e6"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.807188 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ptg5t"] Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.807315 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bbpqw" event={"ID":"5b9d4194-45b6-445b-b882-1982630eaecf","Type":"ContainerStarted","Data":"a8a7f83a2f5a04e287b0f6409cd34cf039a068e910d2cc60d9f697cea2cf5a5d"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.807430 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ptg5t"] Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.807553 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bbpqw" event={"ID":"5b9d4194-45b6-445b-b882-1982630eaecf","Type":"ContainerStarted","Data":"dc590d0bbe0b95a1ecc8ccf0bcd0234f89cf4ec6003b7bea38fe7b664737f621"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.808103 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.813945 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d70a46f0-1e5d-4edb-8745-9b8a41b66241-catalog-content\") pod \"redhat-marketplace-jnw8m\" (UID: \"d70a46f0-1e5d-4edb-8745-9b8a41b66241\") " pod="openshift-marketplace/redhat-marketplace-jnw8m" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.814242 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjjfw\" (UniqueName: \"kubernetes.io/projected/d70a46f0-1e5d-4edb-8745-9b8a41b66241-kube-api-access-fjjfw\") pod \"redhat-marketplace-jnw8m\" (UID: \"d70a46f0-1e5d-4edb-8745-9b8a41b66241\") " pod="openshift-marketplace/redhat-marketplace-jnw8m" Mar 21 08:29:51 crc kubenswrapper[4696]: E0321 08:29:51.814910 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:52.314885205 +0000 UTC m=+126.435765918 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.815089 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d70a46f0-1e5d-4edb-8745-9b8a41b66241-utilities\") pod \"redhat-marketplace-jnw8m\" (UID: \"d70a46f0-1e5d-4edb-8745-9b8a41b66241\") " pod="openshift-marketplace/redhat-marketplace-jnw8m" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.815188 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:51 crc kubenswrapper[4696]: E0321 08:29:51.815470 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:52.31546429 +0000 UTC m=+126.436344993 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.864297 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" event={"ID":"5f37526d-ea7c-4607-920a-ca63aa96184c","Type":"ContainerStarted","Data":"430d7827b29f59b58867e896600db7736b683fcaad9bbd489d978811f63d0ec9"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.865322 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.882959 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-chtnc" event={"ID":"367f9b3a-40fe-414e-bc06-949205afe2b8","Type":"ContainerStarted","Data":"e28ecf175962beea0a31c2f988f623e45b955c4795b46d7c659fab384079a259"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.898000 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4scwl" podStartSLOduration=84.897982277 podStartE2EDuration="1m24.897982277s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:51.874777613 +0000 UTC m=+125.995658346" watchObservedRunningTime="2026-03-21 08:29:51.897982277 +0000 UTC m=+126.018863000" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.909000 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9s5ck" event={"ID":"1df1b2d7-6ec6-4b8c-95d6-d08bdf698974","Type":"ContainerStarted","Data":"72e7bfe1589f712d11a9d46b82d953f7c16629be0e7dedb53564873471aca0ff"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.909784 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9s5ck" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.911271 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" event={"ID":"973fa2b5-c6ce-498f-9b2b-311e8dd34057","Type":"ContainerStarted","Data":"6ed0dd17f07cf887f320e3f6560876112a3577e3274a4b0d84c8e65baede4294"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.913578 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.914999 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bbpqw" podStartSLOduration=84.914982299 podStartE2EDuration="1m24.914982299s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:51.898440978 +0000 UTC m=+126.019321691" watchObservedRunningTime="2026-03-21 08:29:51.914982299 +0000 UTC m=+126.035863012" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.916694 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cfjs8"] Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.921912 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.922150 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjjfw\" (UniqueName: \"kubernetes.io/projected/d70a46f0-1e5d-4edb-8745-9b8a41b66241-kube-api-access-fjjfw\") pod \"redhat-marketplace-jnw8m\" (UID: \"d70a46f0-1e5d-4edb-8745-9b8a41b66241\") " pod="openshift-marketplace/redhat-marketplace-jnw8m" Mar 21 08:29:51 crc kubenswrapper[4696]: E0321 08:29:51.922218 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:52.422197987 +0000 UTC m=+126.543078700 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.922264 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d70a46f0-1e5d-4edb-8745-9b8a41b66241-utilities\") pod \"redhat-marketplace-jnw8m\" (UID: \"d70a46f0-1e5d-4edb-8745-9b8a41b66241\") " pod="openshift-marketplace/redhat-marketplace-jnw8m" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.922344 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.922381 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d70a46f0-1e5d-4edb-8745-9b8a41b66241-catalog-content\") pod \"redhat-marketplace-jnw8m\" (UID: \"d70a46f0-1e5d-4edb-8745-9b8a41b66241\") " pod="openshift-marketplace/redhat-marketplace-jnw8m" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.923522 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d70a46f0-1e5d-4edb-8745-9b8a41b66241-utilities\") pod \"redhat-marketplace-jnw8m\" (UID: \"d70a46f0-1e5d-4edb-8745-9b8a41b66241\") " pod="openshift-marketplace/redhat-marketplace-jnw8m" Mar 21 08:29:51 crc kubenswrapper[4696]: E0321 08:29:51.923735 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:52.423727146 +0000 UTC m=+126.544607859 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.924140 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d70a46f0-1e5d-4edb-8745-9b8a41b66241-catalog-content\") pod \"redhat-marketplace-jnw8m\" (UID: \"d70a46f0-1e5d-4edb-8745-9b8a41b66241\") " pod="openshift-marketplace/redhat-marketplace-jnw8m" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.933539 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" podStartSLOduration=10.933520671 podStartE2EDuration="10.933520671s" podCreationTimestamp="2026-03-21 08:29:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:51.931055357 +0000 UTC m=+126.051936070" watchObservedRunningTime="2026-03-21 08:29:51.933520671 +0000 UTC m=+126.054401384" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.946093 4696 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5vrbx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.946148 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" podUID="973fa2b5-c6ce-498f-9b2b-311e8dd34057" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.947486 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.947514 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2q58n" event={"ID":"aed96ecb-ad63-425d-b8f4-bd55aeb752dd","Type":"ContainerStarted","Data":"96c9e658fced77441f4294604007397ac20608f8de9283951104782d7a7f21da"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.947619 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cfjs8" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.949923 4696 ???:1] "http: TLS handshake error from 192.168.126.11:37486: no serving certificate available for the kubelet" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.954642 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.968531 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjjfw\" (UniqueName: \"kubernetes.io/projected/d70a46f0-1e5d-4edb-8745-9b8a41b66241-kube-api-access-fjjfw\") pod \"redhat-marketplace-jnw8m\" (UID: \"d70a46f0-1e5d-4edb-8745-9b8a41b66241\") " pod="openshift-marketplace/redhat-marketplace-jnw8m" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.970320 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" podStartSLOduration=84.970299848 podStartE2EDuration="1m24.970299848s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:51.963542843 +0000 UTC m=+126.084423556" watchObservedRunningTime="2026-03-21 08:29:51.970299848 +0000 UTC m=+126.091180561" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.972974 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cfjs8"] Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.980737 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9s5ck" Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.988340 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfpb" event={"ID":"ffb878ee-7cd7-4830-bb19-6b2a5e6bd1e0","Type":"ContainerStarted","Data":"a089c6c34b9c2c191e41c1301f286d70c22606ede2608dc6593c336b7de0c4cb"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.988384 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfpb" event={"ID":"ffb878ee-7cd7-4830-bb19-6b2a5e6bd1e0","Type":"ContainerStarted","Data":"2de0345a437f9ee0154273f4167330886e26582ee9f0a2863fe8e78022887613"} Mar 21 08:29:51 crc kubenswrapper[4696]: I0321 08:29:51.989810 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9s5ck" podStartSLOduration=84.989793435 podStartE2EDuration="1m24.989793435s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:51.988865961 +0000 UTC m=+126.109746674" watchObservedRunningTime="2026-03-21 08:29:51.989793435 +0000 UTC m=+126.110674148" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.014384 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8hcm" event={"ID":"b999de84-1fbb-4184-a136-ce68641fc110","Type":"ContainerStarted","Data":"687e7a7a165307dbd7893cc49813063bb4fb8fd5fa08e878ab1df95c8199bbb6"} Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.027874 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.028202 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0030cc8a-f473-4d0a-ace9-728b66e62634-catalog-content\") pod \"redhat-operators-cfjs8\" (UID: \"0030cc8a-f473-4d0a-ace9-728b66e62634\") " pod="openshift-marketplace/redhat-operators-cfjs8" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.028231 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0030cc8a-f473-4d0a-ace9-728b66e62634-utilities\") pod \"redhat-operators-cfjs8\" (UID: \"0030cc8a-f473-4d0a-ace9-728b66e62634\") " pod="openshift-marketplace/redhat-operators-cfjs8" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.028292 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n825s\" (UniqueName: \"kubernetes.io/projected/0030cc8a-f473-4d0a-ace9-728b66e62634-kube-api-access-n825s\") pod \"redhat-operators-cfjs8\" (UID: \"0030cc8a-f473-4d0a-ace9-728b66e62634\") " pod="openshift-marketplace/redhat-operators-cfjs8" Mar 21 08:29:52 crc kubenswrapper[4696]: E0321 08:29:52.029174 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:52.529123399 +0000 UTC m=+126.650004112 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.062151 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfpb" podStartSLOduration=85.062130917 podStartE2EDuration="1m25.062130917s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:52.05725516 +0000 UTC m=+126.178135873" watchObservedRunningTime="2026-03-21 08:29:52.062130917 +0000 UTC m=+126.183011630" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.070510 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jnw8m" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.105627 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj"] Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.109123 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-trvd2"] Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.110566 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-trvd2" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.130927 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n825s\" (UniqueName: \"kubernetes.io/projected/0030cc8a-f473-4d0a-ace9-728b66e62634-kube-api-access-n825s\") pod \"redhat-operators-cfjs8\" (UID: \"0030cc8a-f473-4d0a-ace9-728b66e62634\") " pod="openshift-marketplace/redhat-operators-cfjs8" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.131048 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.131147 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0030cc8a-f473-4d0a-ace9-728b66e62634-catalog-content\") pod \"redhat-operators-cfjs8\" (UID: \"0030cc8a-f473-4d0a-ace9-728b66e62634\") " pod="openshift-marketplace/redhat-operators-cfjs8" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.131177 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0030cc8a-f473-4d0a-ace9-728b66e62634-utilities\") pod \"redhat-operators-cfjs8\" (UID: \"0030cc8a-f473-4d0a-ace9-728b66e62634\") " pod="openshift-marketplace/redhat-operators-cfjs8" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.132671 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0030cc8a-f473-4d0a-ace9-728b66e62634-catalog-content\") pod \"redhat-operators-cfjs8\" (UID: \"0030cc8a-f473-4d0a-ace9-728b66e62634\") " pod="openshift-marketplace/redhat-operators-cfjs8" Mar 21 08:29:52 crc kubenswrapper[4696]: E0321 08:29:52.132725 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:52.632709013 +0000 UTC m=+126.753589726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.135640 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-2q58n" podStartSLOduration=85.135625819 podStartE2EDuration="1m25.135625819s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:52.115570207 +0000 UTC m=+126.236450920" watchObservedRunningTime="2026-03-21 08:29:52.135625819 +0000 UTC m=+126.256506532" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.135862 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-trvd2"] Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.150107 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0030cc8a-f473-4d0a-ace9-728b66e62634-utilities\") pod \"redhat-operators-cfjs8\" (UID: \"0030cc8a-f473-4d0a-ace9-728b66e62634\") " pod="openshift-marketplace/redhat-operators-cfjs8" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.194393 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n825s\" (UniqueName: \"kubernetes.io/projected/0030cc8a-f473-4d0a-ace9-728b66e62634-kube-api-access-n825s\") pod \"redhat-operators-cfjs8\" (UID: \"0030cc8a-f473-4d0a-ace9-728b66e62634\") " pod="openshift-marketplace/redhat-operators-cfjs8" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.221901 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8hcm" podStartSLOduration=85.221881972 podStartE2EDuration="1m25.221881972s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:52.220137067 +0000 UTC m=+126.341017810" watchObservedRunningTime="2026-03-21 08:29:52.221881972 +0000 UTC m=+126.342762685" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.232391 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.232692 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpfnx\" (UniqueName: \"kubernetes.io/projected/d6b6d345-41c3-4926-bf33-bac67ac54755-kube-api-access-dpfnx\") pod \"redhat-marketplace-trvd2\" (UID: \"d6b6d345-41c3-4926-bf33-bac67ac54755\") " pod="openshift-marketplace/redhat-marketplace-trvd2" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.232727 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6b6d345-41c3-4926-bf33-bac67ac54755-catalog-content\") pod \"redhat-marketplace-trvd2\" (UID: \"d6b6d345-41c3-4926-bf33-bac67ac54755\") " pod="openshift-marketplace/redhat-marketplace-trvd2" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.232761 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6b6d345-41c3-4926-bf33-bac67ac54755-utilities\") pod \"redhat-marketplace-trvd2\" (UID: \"d6b6d345-41c3-4926-bf33-bac67ac54755\") " pod="openshift-marketplace/redhat-marketplace-trvd2" Mar 21 08:29:52 crc kubenswrapper[4696]: E0321 08:29:52.232917 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:52.732901369 +0000 UTC m=+126.853782082 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.309562 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jmqr7"] Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.311654 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jmqr7" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.328247 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cfjs8" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.335365 4696 ???:1] "http: TLS handshake error from 192.168.126.11:37496: no serving certificate available for the kubelet" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.348581 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6b6d345-41c3-4926-bf33-bac67ac54755-utilities\") pod \"redhat-marketplace-trvd2\" (UID: \"d6b6d345-41c3-4926-bf33-bac67ac54755\") " pod="openshift-marketplace/redhat-marketplace-trvd2" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.348889 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.349152 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpfnx\" (UniqueName: \"kubernetes.io/projected/d6b6d345-41c3-4926-bf33-bac67ac54755-kube-api-access-dpfnx\") pod \"redhat-marketplace-trvd2\" (UID: \"d6b6d345-41c3-4926-bf33-bac67ac54755\") " pod="openshift-marketplace/redhat-marketplace-trvd2" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.349279 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6b6d345-41c3-4926-bf33-bac67ac54755-catalog-content\") pod \"redhat-marketplace-trvd2\" (UID: \"d6b6d345-41c3-4926-bf33-bac67ac54755\") " pod="openshift-marketplace/redhat-marketplace-trvd2" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.349512 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6b6d345-41c3-4926-bf33-bac67ac54755-utilities\") pod \"redhat-marketplace-trvd2\" (UID: \"d6b6d345-41c3-4926-bf33-bac67ac54755\") " pod="openshift-marketplace/redhat-marketplace-trvd2" Mar 21 08:29:52 crc kubenswrapper[4696]: E0321 08:29:52.349858 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:52.849843352 +0000 UTC m=+126.970724065 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.350272 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6b6d345-41c3-4926-bf33-bac67ac54755-catalog-content\") pod \"redhat-marketplace-trvd2\" (UID: \"d6b6d345-41c3-4926-bf33-bac67ac54755\") " pod="openshift-marketplace/redhat-marketplace-trvd2" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.369888 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jmqr7"] Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.413648 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpfnx\" (UniqueName: \"kubernetes.io/projected/d6b6d345-41c3-4926-bf33-bac67ac54755-kube-api-access-dpfnx\") pod \"redhat-marketplace-trvd2\" (UID: \"d6b6d345-41c3-4926-bf33-bac67ac54755\") " pod="openshift-marketplace/redhat-marketplace-trvd2" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.426857 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-trvd2" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.453812 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.454121 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ed31bad-6fb5-49be-922b-27198867ca4a-utilities\") pod \"redhat-operators-jmqr7\" (UID: \"6ed31bad-6fb5-49be-922b-27198867ca4a\") " pod="openshift-marketplace/redhat-operators-jmqr7" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.454155 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ed31bad-6fb5-49be-922b-27198867ca4a-catalog-content\") pod \"redhat-operators-jmqr7\" (UID: \"6ed31bad-6fb5-49be-922b-27198867ca4a\") " pod="openshift-marketplace/redhat-operators-jmqr7" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.454205 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9q4bs\" (UniqueName: \"kubernetes.io/projected/6ed31bad-6fb5-49be-922b-27198867ca4a-kube-api-access-9q4bs\") pod \"redhat-operators-jmqr7\" (UID: \"6ed31bad-6fb5-49be-922b-27198867ca4a\") " pod="openshift-marketplace/redhat-operators-jmqr7" Mar 21 08:29:52 crc kubenswrapper[4696]: E0321 08:29:52.454290 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:52.954264028 +0000 UTC m=+127.075144741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.454417 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:52 crc kubenswrapper[4696]: E0321 08:29:52.454758 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:52.954751771 +0000 UTC m=+127.075632484 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.556193 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.556785 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ed31bad-6fb5-49be-922b-27198867ca4a-utilities\") pod \"redhat-operators-jmqr7\" (UID: \"6ed31bad-6fb5-49be-922b-27198867ca4a\") " pod="openshift-marketplace/redhat-operators-jmqr7" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.556830 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ed31bad-6fb5-49be-922b-27198867ca4a-catalog-content\") pod \"redhat-operators-jmqr7\" (UID: \"6ed31bad-6fb5-49be-922b-27198867ca4a\") " pod="openshift-marketplace/redhat-operators-jmqr7" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.556890 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9q4bs\" (UniqueName: \"kubernetes.io/projected/6ed31bad-6fb5-49be-922b-27198867ca4a-kube-api-access-9q4bs\") pod \"redhat-operators-jmqr7\" (UID: \"6ed31bad-6fb5-49be-922b-27198867ca4a\") " pod="openshift-marketplace/redhat-operators-jmqr7" Mar 21 08:29:52 crc kubenswrapper[4696]: E0321 08:29:52.557266 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:53.057250717 +0000 UTC m=+127.178131430 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.557878 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ed31bad-6fb5-49be-922b-27198867ca4a-utilities\") pod \"redhat-operators-jmqr7\" (UID: \"6ed31bad-6fb5-49be-922b-27198867ca4a\") " pod="openshift-marketplace/redhat-operators-jmqr7" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.558100 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ed31bad-6fb5-49be-922b-27198867ca4a-catalog-content\") pod \"redhat-operators-jmqr7\" (UID: \"6ed31bad-6fb5-49be-922b-27198867ca4a\") " pod="openshift-marketplace/redhat-operators-jmqr7" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.566808 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c" path="/var/lib/kubelet/pods/0d33a4e5-a970-4e6e-b4c4-87d6b1d9609c/volumes" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.567536 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14fd07c2-17e1-4adb-87c8-459f1dcb41c2" path="/var/lib/kubelet/pods/14fd07c2-17e1-4adb-87c8-459f1dcb41c2/volumes" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.621533 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9q4bs\" (UniqueName: \"kubernetes.io/projected/6ed31bad-6fb5-49be-922b-27198867ca4a-kube-api-access-9q4bs\") pod \"redhat-operators-jmqr7\" (UID: \"6ed31bad-6fb5-49be-922b-27198867ca4a\") " pod="openshift-marketplace/redhat-operators-jmqr7" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.648037 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-wkbtr"] Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.660691 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:52 crc kubenswrapper[4696]: E0321 08:29:52.661007 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:53.160995086 +0000 UTC m=+127.281875799 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.711466 4696 patch_prober.go:28] interesting pod/router-default-5444994796-sdgwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 08:29:52 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Mar 21 08:29:52 crc kubenswrapper[4696]: [+]process-running ok Mar 21 08:29:52 crc kubenswrapper[4696]: healthz check failed Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.711520 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sdgwl" podUID="231fd5a3-3560-4a89-a5be-59880089a8fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.724119 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jmqr7" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.736762 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jnw8m"] Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.761805 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:52 crc kubenswrapper[4696]: E0321 08:29:52.762143 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:53.262126787 +0000 UTC m=+127.383007500 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.863420 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:52 crc kubenswrapper[4696]: E0321 08:29:52.864722 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:53.364706055 +0000 UTC m=+127.485586768 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.875766 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cfjs8"] Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.887747 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj"] Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.888467 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.894877 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.895107 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.895278 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.895386 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 21 08:29:52 crc kubenswrapper[4696]: W0321 08:29:52.895578 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0030cc8a_f473_4d0a_ace9_728b66e62634.slice/crio-08ec3a1b8717c1c8352f9398afd7a2b7497ef06f68318a7ae0c59170c0aeb3b9 WatchSource:0}: Error finding container 08ec3a1b8717c1c8352f9398afd7a2b7497ef06f68318a7ae0c59170c0aeb3b9: Status 404 returned error can't find the container with id 08ec3a1b8717c1c8352f9398afd7a2b7497ef06f68318a7ae0c59170c0aeb3b9 Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.899680 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.900531 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.909254 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.911982 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj"] Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.973255 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.973604 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-config\") pod \"controller-manager-6c7d9ccbdf-9nbwj\" (UID: \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\") " pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.973634 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-proxy-ca-bundles\") pod \"controller-manager-6c7d9ccbdf-9nbwj\" (UID: \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\") " pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.973658 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvchl\" (UniqueName: \"kubernetes.io/projected/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-kube-api-access-rvchl\") pod \"controller-manager-6c7d9ccbdf-9nbwj\" (UID: \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\") " pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.973698 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-client-ca\") pod \"controller-manager-6c7d9ccbdf-9nbwj\" (UID: \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\") " pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" Mar 21 08:29:52 crc kubenswrapper[4696]: I0321 08:29:52.973721 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-serving-cert\") pod \"controller-manager-6c7d9ccbdf-9nbwj\" (UID: \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\") " pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" Mar 21 08:29:52 crc kubenswrapper[4696]: E0321 08:29:52.973849 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:53.473834144 +0000 UTC m=+127.594714857 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.040985 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-w95sn" event={"ID":"9bbfb96a-607b-47df-83df-7c75dd94ad71","Type":"ContainerStarted","Data":"1561d666ba89b1f136629ebb243cad6bc2d0e689c685828c31b42f4374002ba8"} Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.041024 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-w95sn" event={"ID":"9bbfb96a-607b-47df-83df-7c75dd94ad71","Type":"ContainerStarted","Data":"5dfc14ed3f7c80137d5919bb0237535831b378de524433c690efc1d349218538"} Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.073799 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qm5pb" event={"ID":"8a77c875-c8d2-40e5-aabc-4f7ffcd29693","Type":"ContainerStarted","Data":"59031aad26e090d03405a35a178047e0da9c9d475f9cd70feffc30617b933cff"} Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.074529 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qm5pb" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.080588 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-client-ca\") pod \"controller-manager-6c7d9ccbdf-9nbwj\" (UID: \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\") " pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.080639 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-serving-cert\") pod \"controller-manager-6c7d9ccbdf-9nbwj\" (UID: \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\") " pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.080695 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-config\") pod \"controller-manager-6c7d9ccbdf-9nbwj\" (UID: \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\") " pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.080728 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-proxy-ca-bundles\") pod \"controller-manager-6c7d9ccbdf-9nbwj\" (UID: \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\") " pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.080750 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvchl\" (UniqueName: \"kubernetes.io/projected/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-kube-api-access-rvchl\") pod \"controller-manager-6c7d9ccbdf-9nbwj\" (UID: \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\") " pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.080773 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.082209 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-config\") pod \"controller-manager-6c7d9ccbdf-9nbwj\" (UID: \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\") " pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.082977 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-client-ca\") pod \"controller-manager-6c7d9ccbdf-9nbwj\" (UID: \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\") " pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.083158 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-proxy-ca-bundles\") pod \"controller-manager-6c7d9ccbdf-9nbwj\" (UID: \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\") " pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.083682 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wshd6" event={"ID":"9d4f19c5-4861-4ae6-99d0-d32e07de56e2","Type":"ContainerStarted","Data":"3251d75c3d98ef35f0ad40e13f0998799628f8d05cbcda4228fb5e2bb89fd8b5"} Mar 21 08:29:53 crc kubenswrapper[4696]: E0321 08:29:53.084194 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:53.584181365 +0000 UTC m=+127.705062078 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.096997 4696 ???:1] "http: TLS handshake error from 192.168.126.11:37508: no serving certificate available for the kubelet" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.107743 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-serving-cert\") pod \"controller-manager-6c7d9ccbdf-9nbwj\" (UID: \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\") " pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.108186 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-w95sn" podStartSLOduration=87.108175399 podStartE2EDuration="1m27.108175399s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:53.108122028 +0000 UTC m=+127.229002741" watchObservedRunningTime="2026-03-21 08:29:53.108175399 +0000 UTC m=+127.229056112" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.172577 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" event={"ID":"91299f09-692d-4c4e-a83c-22fcd785c1ef","Type":"ContainerStarted","Data":"8f967bae0794d822403dd54c25e20113704807775d470e3abc72627abd85a8fe"} Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.174074 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.181888 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:53 crc kubenswrapper[4696]: E0321 08:29:53.183033 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:53.683015336 +0000 UTC m=+127.803896049 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.186644 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvchl\" (UniqueName: \"kubernetes.io/projected/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-kube-api-access-rvchl\") pod \"controller-manager-6c7d9ccbdf-9nbwj\" (UID: \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\") " pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.189220 4696 generic.go:334] "Generic (PLEG): container finished" podID="85a5000b-3b68-4808-bcfc-93bcbde10dde" containerID="473840c7910cf1720058db29c86564e69cb039cf4db0ba23a177368175ed1f58" exitCode=0 Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.189314 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v9rlg" event={"ID":"85a5000b-3b68-4808-bcfc-93bcbde10dde","Type":"ContainerDied","Data":"473840c7910cf1720058db29c86564e69cb039cf4db0ba23a177368175ed1f58"} Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.189348 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v9rlg" event={"ID":"85a5000b-3b68-4808-bcfc-93bcbde10dde","Type":"ContainerStarted","Data":"404a5067b9555350774dc249e338ff2987f6c02171ba5c83b7c05b916470b110"} Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.205464 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.224574 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.225640 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jnw8m" event={"ID":"d70a46f0-1e5d-4edb-8745-9b8a41b66241","Type":"ContainerStarted","Data":"c421f19fe2b512c1558d4402b703acf79f57365a6fbf5ec4ecb952e7c7904356"} Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.238683 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qm5pb" podStartSLOduration=86.238662063 podStartE2EDuration="1m26.238662063s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:53.158324894 +0000 UTC m=+127.279205607" watchObservedRunningTime="2026-03-21 08:29:53.238662063 +0000 UTC m=+127.359542776" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.240620 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" podStartSLOduration=86.240610494 podStartE2EDuration="1m26.240610494s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:53.230320336 +0000 UTC m=+127.351201049" watchObservedRunningTime="2026-03-21 08:29:53.240610494 +0000 UTC m=+127.361491207" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.272603 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-trvd2"] Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.275075 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-chtnc" event={"ID":"367f9b3a-40fe-414e-bc06-949205afe2b8","Type":"ContainerStarted","Data":"df9e8b788dfa681e99014c7fde8065c60700fd4b5a434a8688f6422840c5b4d8"} Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.275121 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-chtnc" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.283941 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:53 crc kubenswrapper[4696]: E0321 08:29:53.284732 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:53.784721771 +0000 UTC m=+127.905602484 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.288688 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ntmrq" event={"ID":"45a093d5-91a7-44bf-8072-fc24898b79e8","Type":"ContainerStarted","Data":"1e4a685c492a21075b34265e9212bdba2781abbc25fa3a824769db23d1c3d113"} Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.290454 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" event={"ID":"e1690a33-2e30-4c37-ba60-5def18cc3867","Type":"ContainerStarted","Data":"a1e9dd16fb92dbc127e8a88b556a14e21572f31a679145c5a16e034f058e914e"} Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.290478 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" event={"ID":"e1690a33-2e30-4c37-ba60-5def18cc3867","Type":"ContainerStarted","Data":"4d3336bfb1e29e14fab43b72b48a3ca45d88fe9bf1e608647f15c5020407072a"} Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.291196 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.296293 4696 generic.go:334] "Generic (PLEG): container finished" podID="0b88fc0a-dc29-4aa2-b625-c24162c0dc0b" containerID="bd58cdcb5e5ee859ecbd932d8fc0256b43c576c5301c7787d43ff6b4ef9a684b" exitCode=0 Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.296374 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7wvkn" event={"ID":"0b88fc0a-dc29-4aa2-b625-c24162c0dc0b","Type":"ContainerDied","Data":"bd58cdcb5e5ee859ecbd932d8fc0256b43c576c5301c7787d43ff6b4ef9a684b"} Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.296397 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7wvkn" event={"ID":"0b88fc0a-dc29-4aa2-b625-c24162c0dc0b","Type":"ContainerStarted","Data":"89bc54fdd0a2fe3ec8e94d47ce3b3f8409f69e12d70f294b9e6c516115ebbb77"} Mar 21 08:29:53 crc kubenswrapper[4696]: W0321 08:29:53.318166 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6b6d345_41c3_4926_bf33_bac67ac54755.slice/crio-834509e63dca63e1241ecc534865debc65b471600775671608f17370770b6eb0 WatchSource:0}: Error finding container 834509e63dca63e1241ecc534865debc65b471600775671608f17370770b6eb0: Status 404 returned error can't find the container with id 834509e63dca63e1241ecc534865debc65b471600775671608f17370770b6eb0 Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.326528 4696 generic.go:334] "Generic (PLEG): container finished" podID="6aabaca2-7def-4550-bae1-9337b7c1002b" containerID="82b676e04b7e68c50b5598983dadfa70abef0bed245e0d350cb58d9c32cc7f47" exitCode=0 Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.326602 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qsg67" event={"ID":"6aabaca2-7def-4550-bae1-9337b7c1002b","Type":"ContainerDied","Data":"82b676e04b7e68c50b5598983dadfa70abef0bed245e0d350cb58d9c32cc7f47"} Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.326634 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qsg67" event={"ID":"6aabaca2-7def-4550-bae1-9337b7c1002b","Type":"ContainerStarted","Data":"ab7fb5221a42f276201e0f172904818436768bf0492bc00ee90713be326befaf"} Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.398337 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.398417 4696 patch_prober.go:28] interesting pod/console-f9d7485db-f5n2g container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.7:8443/health\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.398472 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-f5n2g" podUID="d1d05966-d105-4c45-bfa2-e66a67882baf" containerName="console" probeResult="failure" output="Get \"https://10.217.0.7:8443/health\": dial tcp 10.217.0.7:8443: connect: connection refused" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.400874 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.401778 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" Mar 21 08:29:53 crc kubenswrapper[4696]: E0321 08:29:53.402743 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:53.902716332 +0000 UTC m=+128.023597045 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.404057 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.406176 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-6m4ch" event={"ID":"e26909f3-14e6-4f5a-9ea9-5b44ac3aa56a","Type":"ContainerStarted","Data":"f3e2aefc250a7e8e2d9b37bee11fd126c2afebf477a1591247c064fb059b382d"} Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.436124 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jmqr7"] Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.459063 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-59mfl" event={"ID":"bf190816-3a09-4918-ab57-d31e55919f3e","Type":"ContainerStarted","Data":"53fecab5432cb948e0812c562ded316f4d4de4579d46753375867fd43bcbcb92"} Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.474892 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" podStartSLOduration=6.474874689 podStartE2EDuration="6.474874689s" podCreationTimestamp="2026-03-21 08:29:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:53.474259073 +0000 UTC m=+127.595139786" watchObservedRunningTime="2026-03-21 08:29:53.474874689 +0000 UTC m=+127.595755402" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.480928 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cfjs8" event={"ID":"0030cc8a-f473-4d0a-ace9-728b66e62634","Type":"ContainerStarted","Data":"08ec3a1b8717c1c8352f9398afd7a2b7497ef06f68318a7ae0c59170c0aeb3b9"} Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.504236 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:53 crc kubenswrapper[4696]: E0321 08:29:53.505023 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:54.005007023 +0000 UTC m=+128.125887736 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.518428 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ntmrq" podStartSLOduration=86.518409701 podStartE2EDuration="1m26.518409701s" podCreationTimestamp="2026-03-21 08:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:53.516001668 +0000 UTC m=+127.636882391" watchObservedRunningTime="2026-03-21 08:29:53.518409701 +0000 UTC m=+127.639290414" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.554138 4696 generic.go:334] "Generic (PLEG): container finished" podID="5149cc5e-d877-4d52-8db6-7df6ae060ab0" containerID="60d54f52b8f62e3f02ae6a78c86e02e67f31dea89ae92fc551f09f878a24b04a" exitCode=0 Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.556121 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7tpgm" event={"ID":"5149cc5e-d877-4d52-8db6-7df6ae060ab0","Type":"ContainerDied","Data":"60d54f52b8f62e3f02ae6a78c86e02e67f31dea89ae92fc551f09f878a24b04a"} Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.556153 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7tpgm" event={"ID":"5149cc5e-d877-4d52-8db6-7df6ae060ab0","Type":"ContainerStarted","Data":"1af20a97d04f7edceaca510842173094d663f9bc603c207f36aa171da3369185"} Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.569928 4696 patch_prober.go:28] interesting pod/downloads-7954f5f757-jdfkn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.569994 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-jdfkn" podUID="f3f5d127-addf-4f3a-a9f0-fcb502db57c0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.590720 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.594250 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-xblff" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.597316 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7fjx" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.604979 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:53 crc kubenswrapper[4696]: E0321 08:29:53.609351 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:54.109314756 +0000 UTC m=+128.230195469 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.627846 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-chtnc" podStartSLOduration=12.627807667 podStartE2EDuration="12.627807667s" podCreationTimestamp="2026-03-21 08:29:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:53.615875587 +0000 UTC m=+127.736756300" watchObservedRunningTime="2026-03-21 08:29:53.627807667 +0000 UTC m=+127.748688380" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.705069 4696 patch_prober.go:28] interesting pod/router-default-5444994796-sdgwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 08:29:53 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Mar 21 08:29:53 crc kubenswrapper[4696]: [+]process-running ok Mar 21 08:29:53 crc kubenswrapper[4696]: healthz check failed Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.705124 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sdgwl" podUID="231fd5a3-3560-4a89-a5be-59880089a8fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.716360 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:53 crc kubenswrapper[4696]: E0321 08:29:53.722209 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:54.222196363 +0000 UTC m=+128.343077076 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.821401 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:53 crc kubenswrapper[4696]: E0321 08:29:53.821782 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:54.321733052 +0000 UTC m=+128.442613765 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.822013 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:53 crc kubenswrapper[4696]: E0321 08:29:53.822329 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:54.322322397 +0000 UTC m=+128.443203110 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.849180 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj"] Mar 21 08:29:53 crc kubenswrapper[4696]: W0321 08:29:53.883101 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3fc5cd50_d5a4_4adf_9ef1_7894814b08c3.slice/crio-ce27886d958a66c78ea7409517fd0e24938be118ca5d08a776a5533031dbe1b9 WatchSource:0}: Error finding container ce27886d958a66c78ea7409517fd0e24938be118ca5d08a776a5533031dbe1b9: Status 404 returned error can't find the container with id ce27886d958a66c78ea7409517fd0e24938be118ca5d08a776a5533031dbe1b9 Mar 21 08:29:53 crc kubenswrapper[4696]: I0321 08:29:53.923320 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:53 crc kubenswrapper[4696]: E0321 08:29:53.923676 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:54.423661524 +0000 UTC m=+128.544542237 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.024552 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:54 crc kubenswrapper[4696]: E0321 08:29:54.025045 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:54.525031881 +0000 UTC m=+128.645912594 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.084385 4696 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.126515 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:54 crc kubenswrapper[4696]: E0321 08:29:54.127073 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:54.627039525 +0000 UTC m=+128.747920238 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.228759 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:54 crc kubenswrapper[4696]: E0321 08:29:54.230358 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:54.730341952 +0000 UTC m=+128.851222665 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.318417 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.318457 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.330969 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:54 crc kubenswrapper[4696]: E0321 08:29:54.331332 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:54.831318069 +0000 UTC m=+128.952198782 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.344996 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.422956 4696 ???:1] "http: TLS handshake error from 192.168.126.11:37522: no serving certificate available for the kubelet" Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.432536 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:54 crc kubenswrapper[4696]: E0321 08:29:54.433525 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:54.93284196 +0000 UTC m=+129.053722673 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.534781 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:54 crc kubenswrapper[4696]: E0321 08:29:54.535273 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:55.035257334 +0000 UTC m=+129.156138047 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.622837 4696 generic.go:334] "Generic (PLEG): container finished" podID="d6b6d345-41c3-4926-bf33-bac67ac54755" containerID="c1faf00a7e367e4b6403ca6d24916fb6f0f207cff786a4717bc6e7d6a0f5c18c" exitCode=0 Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.622889 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trvd2" event={"ID":"d6b6d345-41c3-4926-bf33-bac67ac54755","Type":"ContainerDied","Data":"c1faf00a7e367e4b6403ca6d24916fb6f0f207cff786a4717bc6e7d6a0f5c18c"} Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.622972 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trvd2" event={"ID":"d6b6d345-41c3-4926-bf33-bac67ac54755","Type":"ContainerStarted","Data":"834509e63dca63e1241ecc534865debc65b471600775671608f17370770b6eb0"} Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.643958 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" event={"ID":"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3","Type":"ContainerStarted","Data":"5d3c3d1af92fdcea421c2040343f9f31465c7268a8b2752a6a3923e1fff45744"} Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.644526 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" event={"ID":"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3","Type":"ContainerStarted","Data":"ce27886d958a66c78ea7409517fd0e24938be118ca5d08a776a5533031dbe1b9"} Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.646312 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.654147 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:54 crc kubenswrapper[4696]: E0321 08:29:54.656270 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:55.156252692 +0000 UTC m=+129.277133405 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.670794 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" podStartSLOduration=7.67077196 podStartE2EDuration="7.67077196s" podCreationTimestamp="2026-03-21 08:29:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:54.667598097 +0000 UTC m=+128.788478810" watchObservedRunningTime="2026-03-21 08:29:54.67077196 +0000 UTC m=+128.791652673" Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.671049 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.672295 4696 generic.go:334] "Generic (PLEG): container finished" podID="0030cc8a-f473-4d0a-ace9-728b66e62634" containerID="c7319785d59398994daa2833f5887a3596b380c96b61a56700381ac81bcf2730" exitCode=0 Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.672402 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cfjs8" event={"ID":"0030cc8a-f473-4d0a-ace9-728b66e62634","Type":"ContainerDied","Data":"c7319785d59398994daa2833f5887a3596b380c96b61a56700381ac81bcf2730"} Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.674689 4696 generic.go:334] "Generic (PLEG): container finished" podID="d70a46f0-1e5d-4edb-8745-9b8a41b66241" containerID="8e3278a745c964032c3d9d433afe57aa71f2dcc8c63adac2d0008719bbfb7122" exitCode=0 Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.674787 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jnw8m" event={"ID":"d70a46f0-1e5d-4edb-8745-9b8a41b66241","Type":"ContainerDied","Data":"8e3278a745c964032c3d9d433afe57aa71f2dcc8c63adac2d0008719bbfb7122"} Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.703919 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-sdgwl" Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.710155 4696 patch_prober.go:28] interesting pod/router-default-5444994796-sdgwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 08:29:54 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Mar 21 08:29:54 crc kubenswrapper[4696]: [+]process-running ok Mar 21 08:29:54 crc kubenswrapper[4696]: healthz check failed Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.710207 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sdgwl" podUID="231fd5a3-3560-4a89-a5be-59880089a8fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.714116 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wshd6" event={"ID":"9d4f19c5-4861-4ae6-99d0-d32e07de56e2","Type":"ContainerStarted","Data":"2de83584fe668cdb9afcde565114fe9d54cd355e7db8aa0489c8c3f0336b1a2b"} Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.714174 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wshd6" event={"ID":"9d4f19c5-4861-4ae6-99d0-d32e07de56e2","Type":"ContainerStarted","Data":"2fdb05e800b84bdcfa35fb6752facc607e89a9f208a298b90e0059ff1d362d5b"} Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.714190 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wshd6" event={"ID":"9d4f19c5-4861-4ae6-99d0-d32e07de56e2","Type":"ContainerStarted","Data":"27631df5fc91cce9c91242c275bd03504bb41133c5154fe899f567e8180d20a1"} Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.743108 4696 generic.go:334] "Generic (PLEG): container finished" podID="6ed31bad-6fb5-49be-922b-27198867ca4a" containerID="023cc09213253c72ec64db5ed3785b171de83e56ee907fcd9c47b1d913d3da7c" exitCode=0 Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.744716 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jmqr7" event={"ID":"6ed31bad-6fb5-49be-922b-27198867ca4a","Type":"ContainerDied","Data":"023cc09213253c72ec64db5ed3785b171de83e56ee907fcd9c47b1d913d3da7c"} Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.744746 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jmqr7" event={"ID":"6ed31bad-6fb5-49be-922b-27198867ca4a","Type":"ContainerStarted","Data":"33de0d1956eb6a34068d548181deedcb4c2b44fcc00a13aeb01ed22f8052fbc9"} Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.746544 4696 patch_prober.go:28] interesting pod/downloads-7954f5f757-jdfkn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.746590 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-jdfkn" podUID="f3f5d127-addf-4f3a-a9f0-fcb502db57c0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.749352 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" podUID="5f37526d-ea7c-4607-920a-ca63aa96184c" containerName="kube-multus-additional-cni-plugins" containerID="cri-o://430d7827b29f59b58867e896600db7736b683fcaad9bbd489d978811f63d0ec9" gracePeriod=30 Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.758123 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:54 crc kubenswrapper[4696]: E0321 08:29:54.758263 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:55.258245406 +0000 UTC m=+129.379126119 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.760247 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:54 crc kubenswrapper[4696]: E0321 08:29:54.770978 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-21 08:29:55.270947056 +0000 UTC m=+129.391827849 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c46nr" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.774041 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gdp2l" Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.809730 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-wshd6" podStartSLOduration=13.809710924 podStartE2EDuration="13.809710924s" podCreationTimestamp="2026-03-21 08:29:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:54.80685788 +0000 UTC m=+128.927738623" watchObservedRunningTime="2026-03-21 08:29:54.809710924 +0000 UTC m=+128.930591637" Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.852079 4696 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-21T08:29:54.084418265Z","Handler":null,"Name":""} Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.862210 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:54 crc kubenswrapper[4696]: E0321 08:29:54.862612 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-21 08:29:55.36259574 +0000 UTC m=+129.483476453 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.868016 4696 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.868043 4696 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.915657 4696 patch_prober.go:28] interesting pod/downloads-7954f5f757-jdfkn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.915707 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-jdfkn" podUID="f3f5d127-addf-4f3a-a9f0-fcb502db57c0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.915862 4696 patch_prober.go:28] interesting pod/downloads-7954f5f757-jdfkn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.915909 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-jdfkn" podUID="f3f5d127-addf-4f3a-a9f0-fcb502db57c0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.965545 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.971984 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 08:29:54 crc kubenswrapper[4696]: I0321 08:29:54.972033 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:55 crc kubenswrapper[4696]: I0321 08:29:55.004058 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c46nr\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:55 crc kubenswrapper[4696]: I0321 08:29:55.066893 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 21 08:29:55 crc kubenswrapper[4696]: I0321 08:29:55.108183 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 21 08:29:55 crc kubenswrapper[4696]: E0321 08:29:55.145587 4696 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="430d7827b29f59b58867e896600db7736b683fcaad9bbd489d978811f63d0ec9" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 21 08:29:55 crc kubenswrapper[4696]: I0321 08:29:55.150715 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 21 08:29:55 crc kubenswrapper[4696]: I0321 08:29:55.157286 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:55 crc kubenswrapper[4696]: E0321 08:29:55.166664 4696 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="430d7827b29f59b58867e896600db7736b683fcaad9bbd489d978811f63d0ec9" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 21 08:29:55 crc kubenswrapper[4696]: E0321 08:29:55.174990 4696 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="430d7827b29f59b58867e896600db7736b683fcaad9bbd489d978811f63d0ec9" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 21 08:29:55 crc kubenswrapper[4696]: E0321 08:29:55.175081 4696 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" podUID="5f37526d-ea7c-4607-920a-ca63aa96184c" containerName="kube-multus-additional-cni-plugins" Mar 21 08:29:55 crc kubenswrapper[4696]: I0321 08:29:55.702870 4696 patch_prober.go:28] interesting pod/router-default-5444994796-sdgwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 08:29:55 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Mar 21 08:29:55 crc kubenswrapper[4696]: [+]process-running ok Mar 21 08:29:55 crc kubenswrapper[4696]: healthz check failed Mar 21 08:29:55 crc kubenswrapper[4696]: I0321 08:29:55.702945 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sdgwl" podUID="231fd5a3-3560-4a89-a5be-59880089a8fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 08:29:55 crc kubenswrapper[4696]: I0321 08:29:55.769166 4696 generic.go:334] "Generic (PLEG): container finished" podID="6ef60c27-2062-4bc8-8c37-e13db97d29a3" containerID="26fd12ef9eb9bd83c4e74724cce711f57f7190be8e20833ce132ed584b0c3116" exitCode=0 Mar 21 08:29:55 crc kubenswrapper[4696]: I0321 08:29:55.769871 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48" event={"ID":"6ef60c27-2062-4bc8-8c37-e13db97d29a3","Type":"ContainerDied","Data":"26fd12ef9eb9bd83c4e74724cce711f57f7190be8e20833ce132ed584b0c3116"} Mar 21 08:29:55 crc kubenswrapper[4696]: I0321 08:29:55.786367 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c46nr"] Mar 21 08:29:55 crc kubenswrapper[4696]: I0321 08:29:55.820597 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 21 08:29:55 crc kubenswrapper[4696]: I0321 08:29:55.821617 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 21 08:29:55 crc kubenswrapper[4696]: I0321 08:29:55.826493 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 21 08:29:55 crc kubenswrapper[4696]: I0321 08:29:55.829872 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 21 08:29:55 crc kubenswrapper[4696]: I0321 08:29:55.830140 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 21 08:29:55 crc kubenswrapper[4696]: I0321 08:29:55.894191 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6836f01-7f9f-4cc7-99d8-10eababb2b1f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f6836f01-7f9f-4cc7-99d8-10eababb2b1f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 21 08:29:55 crc kubenswrapper[4696]: I0321 08:29:55.894294 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f6836f01-7f9f-4cc7-99d8-10eababb2b1f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f6836f01-7f9f-4cc7-99d8-10eababb2b1f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 21 08:29:55 crc kubenswrapper[4696]: I0321 08:29:55.905572 4696 ???:1] "http: TLS handshake error from 192.168.126.11:37524: no serving certificate available for the kubelet" Mar 21 08:29:56 crc kubenswrapper[4696]: I0321 08:29:56.001455 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6836f01-7f9f-4cc7-99d8-10eababb2b1f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f6836f01-7f9f-4cc7-99d8-10eababb2b1f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 21 08:29:56 crc kubenswrapper[4696]: I0321 08:29:56.001612 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f6836f01-7f9f-4cc7-99d8-10eababb2b1f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f6836f01-7f9f-4cc7-99d8-10eababb2b1f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 21 08:29:56 crc kubenswrapper[4696]: I0321 08:29:56.001716 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f6836f01-7f9f-4cc7-99d8-10eababb2b1f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f6836f01-7f9f-4cc7-99d8-10eababb2b1f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 21 08:29:56 crc kubenswrapper[4696]: I0321 08:29:56.046320 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6836f01-7f9f-4cc7-99d8-10eababb2b1f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f6836f01-7f9f-4cc7-99d8-10eababb2b1f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 21 08:29:56 crc kubenswrapper[4696]: I0321 08:29:56.160449 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 21 08:29:56 crc kubenswrapper[4696]: I0321 08:29:56.575468 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 21 08:29:56 crc kubenswrapper[4696]: I0321 08:29:56.681523 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 21 08:29:56 crc kubenswrapper[4696]: I0321 08:29:56.705228 4696 patch_prober.go:28] interesting pod/router-default-5444994796-sdgwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 08:29:56 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Mar 21 08:29:56 crc kubenswrapper[4696]: [+]process-running ok Mar 21 08:29:56 crc kubenswrapper[4696]: healthz check failed Mar 21 08:29:56 crc kubenswrapper[4696]: I0321 08:29:56.705287 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sdgwl" podUID="231fd5a3-3560-4a89-a5be-59880089a8fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 08:29:56 crc kubenswrapper[4696]: W0321 08:29:56.746770 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podf6836f01_7f9f_4cc7_99d8_10eababb2b1f.slice/crio-360dd776a90671e1b85f2211954dc941129696216a90af58a8804ae8ccc55bcf WatchSource:0}: Error finding container 360dd776a90671e1b85f2211954dc941129696216a90af58a8804ae8ccc55bcf: Status 404 returned error can't find the container with id 360dd776a90671e1b85f2211954dc941129696216a90af58a8804ae8ccc55bcf Mar 21 08:29:56 crc kubenswrapper[4696]: I0321 08:29:56.782907 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f6836f01-7f9f-4cc7-99d8-10eababb2b1f","Type":"ContainerStarted","Data":"360dd776a90671e1b85f2211954dc941129696216a90af58a8804ae8ccc55bcf"} Mar 21 08:29:56 crc kubenswrapper[4696]: I0321 08:29:56.786560 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" event={"ID":"96b62643-30a8-47d5-8fe3-20eadf06e0f7","Type":"ContainerStarted","Data":"604fb7efc7658349c2c84a277fbbb4ba36f5ca4ba8e34e60ccebde2f771fc479"} Mar 21 08:29:56 crc kubenswrapper[4696]: I0321 08:29:56.786587 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" event={"ID":"96b62643-30a8-47d5-8fe3-20eadf06e0f7","Type":"ContainerStarted","Data":"f8fa3fbd51c5a21675e4119072861913eca893ce0c94a48c6e8092e25374c888"} Mar 21 08:29:56 crc kubenswrapper[4696]: I0321 08:29:56.787075 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:29:56 crc kubenswrapper[4696]: I0321 08:29:56.807792 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" podStartSLOduration=90.807772893 podStartE2EDuration="1m30.807772893s" podCreationTimestamp="2026-03-21 08:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:56.804704793 +0000 UTC m=+130.925585506" watchObservedRunningTime="2026-03-21 08:29:56.807772893 +0000 UTC m=+130.928653606" Mar 21 08:29:57 crc kubenswrapper[4696]: I0321 08:29:57.059493 4696 ???:1] "http: TLS handshake error from 192.168.126.11:40034: no serving certificate available for the kubelet" Mar 21 08:29:57 crc kubenswrapper[4696]: I0321 08:29:57.303621 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48" Mar 21 08:29:57 crc kubenswrapper[4696]: I0321 08:29:57.464109 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hbtq\" (UniqueName: \"kubernetes.io/projected/6ef60c27-2062-4bc8-8c37-e13db97d29a3-kube-api-access-8hbtq\") pod \"6ef60c27-2062-4bc8-8c37-e13db97d29a3\" (UID: \"6ef60c27-2062-4bc8-8c37-e13db97d29a3\") " Mar 21 08:29:57 crc kubenswrapper[4696]: I0321 08:29:57.464178 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6ef60c27-2062-4bc8-8c37-e13db97d29a3-secret-volume\") pod \"6ef60c27-2062-4bc8-8c37-e13db97d29a3\" (UID: \"6ef60c27-2062-4bc8-8c37-e13db97d29a3\") " Mar 21 08:29:57 crc kubenswrapper[4696]: I0321 08:29:57.464265 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6ef60c27-2062-4bc8-8c37-e13db97d29a3-config-volume\") pod \"6ef60c27-2062-4bc8-8c37-e13db97d29a3\" (UID: \"6ef60c27-2062-4bc8-8c37-e13db97d29a3\") " Mar 21 08:29:57 crc kubenswrapper[4696]: I0321 08:29:57.465470 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ef60c27-2062-4bc8-8c37-e13db97d29a3-config-volume" (OuterVolumeSpecName: "config-volume") pod "6ef60c27-2062-4bc8-8c37-e13db97d29a3" (UID: "6ef60c27-2062-4bc8-8c37-e13db97d29a3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:29:57 crc kubenswrapper[4696]: I0321 08:29:57.471312 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ef60c27-2062-4bc8-8c37-e13db97d29a3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6ef60c27-2062-4bc8-8c37-e13db97d29a3" (UID: "6ef60c27-2062-4bc8-8c37-e13db97d29a3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:29:57 crc kubenswrapper[4696]: I0321 08:29:57.472420 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ef60c27-2062-4bc8-8c37-e13db97d29a3-kube-api-access-8hbtq" (OuterVolumeSpecName: "kube-api-access-8hbtq") pod "6ef60c27-2062-4bc8-8c37-e13db97d29a3" (UID: "6ef60c27-2062-4bc8-8c37-e13db97d29a3"). InnerVolumeSpecName "kube-api-access-8hbtq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:29:57 crc kubenswrapper[4696]: I0321 08:29:57.545883 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 21 08:29:57 crc kubenswrapper[4696]: I0321 08:29:57.565663 4696 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6ef60c27-2062-4bc8-8c37-e13db97d29a3-config-volume\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:57 crc kubenswrapper[4696]: I0321 08:29:57.565709 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hbtq\" (UniqueName: \"kubernetes.io/projected/6ef60c27-2062-4bc8-8c37-e13db97d29a3-kube-api-access-8hbtq\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:57 crc kubenswrapper[4696]: I0321 08:29:57.565725 4696 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6ef60c27-2062-4bc8-8c37-e13db97d29a3-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 21 08:29:57 crc kubenswrapper[4696]: I0321 08:29:57.703026 4696 patch_prober.go:28] interesting pod/router-default-5444994796-sdgwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 08:29:57 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Mar 21 08:29:57 crc kubenswrapper[4696]: [+]process-running ok Mar 21 08:29:57 crc kubenswrapper[4696]: healthz check failed Mar 21 08:29:57 crc kubenswrapper[4696]: I0321 08:29:57.703099 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sdgwl" podUID="231fd5a3-3560-4a89-a5be-59880089a8fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 08:29:57 crc kubenswrapper[4696]: I0321 08:29:57.812575 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f6836f01-7f9f-4cc7-99d8-10eababb2b1f","Type":"ContainerStarted","Data":"59d70938b4c9fd3a6808891e01e15658045d123d37409e08f10c300f19b04d6e"} Mar 21 08:29:57 crc kubenswrapper[4696]: I0321 08:29:57.825314 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48" Mar 21 08:29:57 crc kubenswrapper[4696]: I0321 08:29:57.826245 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48" event={"ID":"6ef60c27-2062-4bc8-8c37-e13db97d29a3","Type":"ContainerDied","Data":"5a46b0e020a5fc8eeada2c072e338cc86a14724276bce439f8b313ba08e6bac6"} Mar 21 08:29:57 crc kubenswrapper[4696]: I0321 08:29:57.826297 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a46b0e020a5fc8eeada2c072e338cc86a14724276bce439f8b313ba08e6bac6" Mar 21 08:29:57 crc kubenswrapper[4696]: I0321 08:29:57.832462 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.832448537 podStartE2EDuration="2.832448537s" podCreationTimestamp="2026-03-21 08:29:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:57.826760586 +0000 UTC m=+131.947641289" watchObservedRunningTime="2026-03-21 08:29:57.832448537 +0000 UTC m=+131.953329250" Mar 21 08:29:57 crc kubenswrapper[4696]: I0321 08:29:57.843369 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=0.843355308 podStartE2EDuration="843.355308ms" podCreationTimestamp="2026-03-21 08:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:29:57.841723314 +0000 UTC m=+131.962604037" watchObservedRunningTime="2026-03-21 08:29:57.843355308 +0000 UTC m=+131.964236021" Mar 21 08:29:58 crc kubenswrapper[4696]: I0321 08:29:58.702382 4696 patch_prober.go:28] interesting pod/router-default-5444994796-sdgwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 08:29:58 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Mar 21 08:29:58 crc kubenswrapper[4696]: [+]process-running ok Mar 21 08:29:58 crc kubenswrapper[4696]: healthz check failed Mar 21 08:29:58 crc kubenswrapper[4696]: I0321 08:29:58.702445 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sdgwl" podUID="231fd5a3-3560-4a89-a5be-59880089a8fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 08:29:58 crc kubenswrapper[4696]: I0321 08:29:58.792365 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 21 08:29:58 crc kubenswrapper[4696]: E0321 08:29:58.792879 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ef60c27-2062-4bc8-8c37-e13db97d29a3" containerName="collect-profiles" Mar 21 08:29:58 crc kubenswrapper[4696]: I0321 08:29:58.792898 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ef60c27-2062-4bc8-8c37-e13db97d29a3" containerName="collect-profiles" Mar 21 08:29:58 crc kubenswrapper[4696]: I0321 08:29:58.793006 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ef60c27-2062-4bc8-8c37-e13db97d29a3" containerName="collect-profiles" Mar 21 08:29:58 crc kubenswrapper[4696]: I0321 08:29:58.793462 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 21 08:29:58 crc kubenswrapper[4696]: I0321 08:29:58.796280 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 21 08:29:58 crc kubenswrapper[4696]: I0321 08:29:58.796607 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 21 08:29:58 crc kubenswrapper[4696]: I0321 08:29:58.801155 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 21 08:29:58 crc kubenswrapper[4696]: I0321 08:29:58.844766 4696 generic.go:334] "Generic (PLEG): container finished" podID="f6836f01-7f9f-4cc7-99d8-10eababb2b1f" containerID="59d70938b4c9fd3a6808891e01e15658045d123d37409e08f10c300f19b04d6e" exitCode=0 Mar 21 08:29:58 crc kubenswrapper[4696]: I0321 08:29:58.844848 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f6836f01-7f9f-4cc7-99d8-10eababb2b1f","Type":"ContainerDied","Data":"59d70938b4c9fd3a6808891e01e15658045d123d37409e08f10c300f19b04d6e"} Mar 21 08:29:58 crc kubenswrapper[4696]: I0321 08:29:58.895335 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c14af961-f06f-4759-aff6-2b69ec8914e0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c14af961-f06f-4759-aff6-2b69ec8914e0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 21 08:29:58 crc kubenswrapper[4696]: I0321 08:29:58.895379 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c14af961-f06f-4759-aff6-2b69ec8914e0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c14af961-f06f-4759-aff6-2b69ec8914e0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 21 08:29:58 crc kubenswrapper[4696]: I0321 08:29:58.997142 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c14af961-f06f-4759-aff6-2b69ec8914e0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c14af961-f06f-4759-aff6-2b69ec8914e0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 21 08:29:58 crc kubenswrapper[4696]: I0321 08:29:58.997185 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c14af961-f06f-4759-aff6-2b69ec8914e0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c14af961-f06f-4759-aff6-2b69ec8914e0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 21 08:29:58 crc kubenswrapper[4696]: I0321 08:29:58.997257 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c14af961-f06f-4759-aff6-2b69ec8914e0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c14af961-f06f-4759-aff6-2b69ec8914e0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 21 08:29:59 crc kubenswrapper[4696]: I0321 08:29:59.016617 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c14af961-f06f-4759-aff6-2b69ec8914e0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c14af961-f06f-4759-aff6-2b69ec8914e0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 21 08:29:59 crc kubenswrapper[4696]: I0321 08:29:59.113458 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 21 08:29:59 crc kubenswrapper[4696]: I0321 08:29:59.702241 4696 patch_prober.go:28] interesting pod/router-default-5444994796-sdgwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 08:29:59 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Mar 21 08:29:59 crc kubenswrapper[4696]: [+]process-running ok Mar 21 08:29:59 crc kubenswrapper[4696]: healthz check failed Mar 21 08:29:59 crc kubenswrapper[4696]: I0321 08:29:59.702958 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sdgwl" podUID="231fd5a3-3560-4a89-a5be-59880089a8fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 08:29:59 crc kubenswrapper[4696]: I0321 08:29:59.747448 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 21 08:29:59 crc kubenswrapper[4696]: W0321 08:29:59.788080 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podc14af961_f06f_4759_aff6_2b69ec8914e0.slice/crio-a7a75348e92cefc3fb6fa6439ebf1aefe4f7d76bd1f3365fa335edeeb72d94cd WatchSource:0}: Error finding container a7a75348e92cefc3fb6fa6439ebf1aefe4f7d76bd1f3365fa335edeeb72d94cd: Status 404 returned error can't find the container with id a7a75348e92cefc3fb6fa6439ebf1aefe4f7d76bd1f3365fa335edeeb72d94cd Mar 21 08:29:59 crc kubenswrapper[4696]: I0321 08:29:59.903292 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c14af961-f06f-4759-aff6-2b69ec8914e0","Type":"ContainerStarted","Data":"a7a75348e92cefc3fb6fa6439ebf1aefe4f7d76bd1f3365fa335edeeb72d94cd"} Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.090371 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-chtnc" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.141644 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568030-4x55c"] Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.142882 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568030-4x55c" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.145918 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.146124 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.146236 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.146319 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568030-n2gf7"] Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.147060 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568030-n2gf7" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.148125 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.148367 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.174294 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568030-4x55c"] Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.174341 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568030-n2gf7"] Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.237993 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgqbs\" (UniqueName: \"kubernetes.io/projected/a6581056-7677-491f-8f34-5626b43b6e77-kube-api-access-bgqbs\") pod \"auto-csr-approver-29568030-4x55c\" (UID: \"a6581056-7677-491f-8f34-5626b43b6e77\") " pod="openshift-infra/auto-csr-approver-29568030-4x55c" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.238156 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df64449e-f8aa-4e41-9b54-1dcc659d995c-secret-volume\") pod \"collect-profiles-29568030-n2gf7\" (UID: \"df64449e-f8aa-4e41-9b54-1dcc659d995c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568030-n2gf7" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.238273 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prwvn\" (UniqueName: \"kubernetes.io/projected/df64449e-f8aa-4e41-9b54-1dcc659d995c-kube-api-access-prwvn\") pod \"collect-profiles-29568030-n2gf7\" (UID: \"df64449e-f8aa-4e41-9b54-1dcc659d995c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568030-n2gf7" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.238313 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df64449e-f8aa-4e41-9b54-1dcc659d995c-config-volume\") pod \"collect-profiles-29568030-n2gf7\" (UID: \"df64449e-f8aa-4e41-9b54-1dcc659d995c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568030-n2gf7" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.244849 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.339808 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f6836f01-7f9f-4cc7-99d8-10eababb2b1f-kubelet-dir\") pod \"f6836f01-7f9f-4cc7-99d8-10eababb2b1f\" (UID: \"f6836f01-7f9f-4cc7-99d8-10eababb2b1f\") " Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.340160 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6836f01-7f9f-4cc7-99d8-10eababb2b1f-kube-api-access\") pod \"f6836f01-7f9f-4cc7-99d8-10eababb2b1f\" (UID: \"f6836f01-7f9f-4cc7-99d8-10eababb2b1f\") " Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.340186 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f6836f01-7f9f-4cc7-99d8-10eababb2b1f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f6836f01-7f9f-4cc7-99d8-10eababb2b1f" (UID: "f6836f01-7f9f-4cc7-99d8-10eababb2b1f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.340384 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df64449e-f8aa-4e41-9b54-1dcc659d995c-secret-volume\") pod \"collect-profiles-29568030-n2gf7\" (UID: \"df64449e-f8aa-4e41-9b54-1dcc659d995c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568030-n2gf7" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.340434 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prwvn\" (UniqueName: \"kubernetes.io/projected/df64449e-f8aa-4e41-9b54-1dcc659d995c-kube-api-access-prwvn\") pod \"collect-profiles-29568030-n2gf7\" (UID: \"df64449e-f8aa-4e41-9b54-1dcc659d995c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568030-n2gf7" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.340457 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df64449e-f8aa-4e41-9b54-1dcc659d995c-config-volume\") pod \"collect-profiles-29568030-n2gf7\" (UID: \"df64449e-f8aa-4e41-9b54-1dcc659d995c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568030-n2gf7" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.340480 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgqbs\" (UniqueName: \"kubernetes.io/projected/a6581056-7677-491f-8f34-5626b43b6e77-kube-api-access-bgqbs\") pod \"auto-csr-approver-29568030-4x55c\" (UID: \"a6581056-7677-491f-8f34-5626b43b6e77\") " pod="openshift-infra/auto-csr-approver-29568030-4x55c" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.340523 4696 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f6836f01-7f9f-4cc7-99d8-10eababb2b1f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.341999 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df64449e-f8aa-4e41-9b54-1dcc659d995c-config-volume\") pod \"collect-profiles-29568030-n2gf7\" (UID: \"df64449e-f8aa-4e41-9b54-1dcc659d995c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568030-n2gf7" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.356098 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prwvn\" (UniqueName: \"kubernetes.io/projected/df64449e-f8aa-4e41-9b54-1dcc659d995c-kube-api-access-prwvn\") pod \"collect-profiles-29568030-n2gf7\" (UID: \"df64449e-f8aa-4e41-9b54-1dcc659d995c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568030-n2gf7" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.359041 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6836f01-7f9f-4cc7-99d8-10eababb2b1f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f6836f01-7f9f-4cc7-99d8-10eababb2b1f" (UID: "f6836f01-7f9f-4cc7-99d8-10eababb2b1f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.359249 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgqbs\" (UniqueName: \"kubernetes.io/projected/a6581056-7677-491f-8f34-5626b43b6e77-kube-api-access-bgqbs\") pod \"auto-csr-approver-29568030-4x55c\" (UID: \"a6581056-7677-491f-8f34-5626b43b6e77\") " pod="openshift-infra/auto-csr-approver-29568030-4x55c" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.363472 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df64449e-f8aa-4e41-9b54-1dcc659d995c-secret-volume\") pod \"collect-profiles-29568030-n2gf7\" (UID: \"df64449e-f8aa-4e41-9b54-1dcc659d995c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568030-n2gf7" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.442480 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6836f01-7f9f-4cc7-99d8-10eababb2b1f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.459213 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.508217 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568030-4x55c" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.522176 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568030-n2gf7" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.704114 4696 patch_prober.go:28] interesting pod/router-default-5444994796-sdgwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 08:30:00 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Mar 21 08:30:00 crc kubenswrapper[4696]: [+]process-running ok Mar 21 08:30:00 crc kubenswrapper[4696]: healthz check failed Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.704163 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sdgwl" podUID="231fd5a3-3560-4a89-a5be-59880089a8fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.884959 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568030-n2gf7"] Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.923500 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f6836f01-7f9f-4cc7-99d8-10eababb2b1f","Type":"ContainerDied","Data":"360dd776a90671e1b85f2211954dc941129696216a90af58a8804ae8ccc55bcf"} Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.923546 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="360dd776a90671e1b85f2211954dc941129696216a90af58a8804ae8ccc55bcf" Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.923624 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 21 08:30:00 crc kubenswrapper[4696]: W0321 08:30:00.924045 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf64449e_f8aa_4e41_9b54_1dcc659d995c.slice/crio-2adf27cb8ae27e4af384395ee676b50a2d50b0539e4df3a753f1048e2fee98ed WatchSource:0}: Error finding container 2adf27cb8ae27e4af384395ee676b50a2d50b0539e4df3a753f1048e2fee98ed: Status 404 returned error can't find the container with id 2adf27cb8ae27e4af384395ee676b50a2d50b0539e4df3a753f1048e2fee98ed Mar 21 08:30:00 crc kubenswrapper[4696]: I0321 08:30:00.948169 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568030-4x55c"] Mar 21 08:30:00 crc kubenswrapper[4696]: W0321 08:30:00.951539 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6581056_7677_491f_8f34_5626b43b6e77.slice/crio-13e173b4a2a99a0f62501dbfef5be4707fb0e0fc4b93c42093c8580e735872e9 WatchSource:0}: Error finding container 13e173b4a2a99a0f62501dbfef5be4707fb0e0fc4b93c42093c8580e735872e9: Status 404 returned error can't find the container with id 13e173b4a2a99a0f62501dbfef5be4707fb0e0fc4b93c42093c8580e735872e9 Mar 21 08:30:01 crc kubenswrapper[4696]: I0321 08:30:01.702445 4696 patch_prober.go:28] interesting pod/router-default-5444994796-sdgwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 08:30:01 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Mar 21 08:30:01 crc kubenswrapper[4696]: [+]process-running ok Mar 21 08:30:01 crc kubenswrapper[4696]: healthz check failed Mar 21 08:30:01 crc kubenswrapper[4696]: I0321 08:30:01.702764 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sdgwl" podUID="231fd5a3-3560-4a89-a5be-59880089a8fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 08:30:01 crc kubenswrapper[4696]: I0321 08:30:01.935166 4696 generic.go:334] "Generic (PLEG): container finished" podID="c14af961-f06f-4759-aff6-2b69ec8914e0" containerID="e85ff3b13646e37132150beb49ca2cba3d94fe1c04b0ccb6bbcb6567ac192901" exitCode=0 Mar 21 08:30:01 crc kubenswrapper[4696]: I0321 08:30:01.935241 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c14af961-f06f-4759-aff6-2b69ec8914e0","Type":"ContainerDied","Data":"e85ff3b13646e37132150beb49ca2cba3d94fe1c04b0ccb6bbcb6567ac192901"} Mar 21 08:30:01 crc kubenswrapper[4696]: I0321 08:30:01.939196 4696 generic.go:334] "Generic (PLEG): container finished" podID="df64449e-f8aa-4e41-9b54-1dcc659d995c" containerID="85895d626856496826a9a8a44f692f0479e3f5cb33877cf1fccfb8cdc3892201" exitCode=0 Mar 21 08:30:01 crc kubenswrapper[4696]: I0321 08:30:01.939233 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568030-n2gf7" event={"ID":"df64449e-f8aa-4e41-9b54-1dcc659d995c","Type":"ContainerDied","Data":"85895d626856496826a9a8a44f692f0479e3f5cb33877cf1fccfb8cdc3892201"} Mar 21 08:30:01 crc kubenswrapper[4696]: I0321 08:30:01.939262 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568030-n2gf7" event={"ID":"df64449e-f8aa-4e41-9b54-1dcc659d995c","Type":"ContainerStarted","Data":"2adf27cb8ae27e4af384395ee676b50a2d50b0539e4df3a753f1048e2fee98ed"} Mar 21 08:30:01 crc kubenswrapper[4696]: I0321 08:30:01.940571 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568030-4x55c" event={"ID":"a6581056-7677-491f-8f34-5626b43b6e77","Type":"ContainerStarted","Data":"13e173b4a2a99a0f62501dbfef5be4707fb0e0fc4b93c42093c8580e735872e9"} Mar 21 08:30:02 crc kubenswrapper[4696]: I0321 08:30:02.701462 4696 patch_prober.go:28] interesting pod/router-default-5444994796-sdgwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 08:30:02 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Mar 21 08:30:02 crc kubenswrapper[4696]: [+]process-running ok Mar 21 08:30:02 crc kubenswrapper[4696]: healthz check failed Mar 21 08:30:02 crc kubenswrapper[4696]: I0321 08:30:02.701535 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sdgwl" podUID="231fd5a3-3560-4a89-a5be-59880089a8fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 08:30:03 crc kubenswrapper[4696]: I0321 08:30:03.366605 4696 patch_prober.go:28] interesting pod/console-f9d7485db-f5n2g container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.7:8443/health\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Mar 21 08:30:03 crc kubenswrapper[4696]: I0321 08:30:03.366679 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-f5n2g" podUID="d1d05966-d105-4c45-bfa2-e66a67882baf" containerName="console" probeResult="failure" output="Get \"https://10.217.0.7:8443/health\": dial tcp 10.217.0.7:8443: connect: connection refused" Mar 21 08:30:03 crc kubenswrapper[4696]: I0321 08:30:03.543855 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 21 08:30:03 crc kubenswrapper[4696]: I0321 08:30:03.702289 4696 patch_prober.go:28] interesting pod/router-default-5444994796-sdgwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 08:30:03 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Mar 21 08:30:03 crc kubenswrapper[4696]: [+]process-running ok Mar 21 08:30:03 crc kubenswrapper[4696]: healthz check failed Mar 21 08:30:03 crc kubenswrapper[4696]: I0321 08:30:03.702531 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sdgwl" podUID="231fd5a3-3560-4a89-a5be-59880089a8fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 08:30:04 crc kubenswrapper[4696]: I0321 08:30:04.701394 4696 patch_prober.go:28] interesting pod/router-default-5444994796-sdgwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 08:30:04 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Mar 21 08:30:04 crc kubenswrapper[4696]: [+]process-running ok Mar 21 08:30:04 crc kubenswrapper[4696]: healthz check failed Mar 21 08:30:04 crc kubenswrapper[4696]: I0321 08:30:04.701445 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sdgwl" podUID="231fd5a3-3560-4a89-a5be-59880089a8fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 08:30:04 crc kubenswrapper[4696]: I0321 08:30:04.914993 4696 patch_prober.go:28] interesting pod/downloads-7954f5f757-jdfkn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Mar 21 08:30:04 crc kubenswrapper[4696]: I0321 08:30:04.915011 4696 patch_prober.go:28] interesting pod/downloads-7954f5f757-jdfkn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Mar 21 08:30:04 crc kubenswrapper[4696]: I0321 08:30:04.915063 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-jdfkn" podUID="f3f5d127-addf-4f3a-a9f0-fcb502db57c0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Mar 21 08:30:04 crc kubenswrapper[4696]: I0321 08:30:04.915070 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-jdfkn" podUID="f3f5d127-addf-4f3a-a9f0-fcb502db57c0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Mar 21 08:30:05 crc kubenswrapper[4696]: E0321 08:30:05.138619 4696 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="430d7827b29f59b58867e896600db7736b683fcaad9bbd489d978811f63d0ec9" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 21 08:30:05 crc kubenswrapper[4696]: E0321 08:30:05.140626 4696 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="430d7827b29f59b58867e896600db7736b683fcaad9bbd489d978811f63d0ec9" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 21 08:30:05 crc kubenswrapper[4696]: E0321 08:30:05.144253 4696 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="430d7827b29f59b58867e896600db7736b683fcaad9bbd489d978811f63d0ec9" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 21 08:30:05 crc kubenswrapper[4696]: E0321 08:30:05.144305 4696 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" podUID="5f37526d-ea7c-4607-920a-ca63aa96184c" containerName="kube-multus-additional-cni-plugins" Mar 21 08:30:05 crc kubenswrapper[4696]: I0321 08:30:05.482674 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj"] Mar 21 08:30:05 crc kubenswrapper[4696]: I0321 08:30:05.482897 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" podUID="3fc5cd50-d5a4-4adf-9ef1-7894814b08c3" containerName="controller-manager" containerID="cri-o://5d3c3d1af92fdcea421c2040343f9f31465c7268a8b2752a6a3923e1fff45744" gracePeriod=30 Mar 21 08:30:05 crc kubenswrapper[4696]: I0321 08:30:05.489736 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj"] Mar 21 08:30:05 crc kubenswrapper[4696]: I0321 08:30:05.490060 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" podUID="e1690a33-2e30-4c37-ba60-5def18cc3867" containerName="route-controller-manager" containerID="cri-o://a1e9dd16fb92dbc127e8a88b556a14e21572f31a679145c5a16e034f058e914e" gracePeriod=30 Mar 21 08:30:05 crc kubenswrapper[4696]: I0321 08:30:05.501732 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=2.501713078 podStartE2EDuration="2.501713078s" podCreationTimestamp="2026-03-21 08:30:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:30:05.499400516 +0000 UTC m=+139.620281229" watchObservedRunningTime="2026-03-21 08:30:05.501713078 +0000 UTC m=+139.622593791" Mar 21 08:30:05 crc kubenswrapper[4696]: I0321 08:30:05.701675 4696 patch_prober.go:28] interesting pod/router-default-5444994796-sdgwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 08:30:05 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Mar 21 08:30:05 crc kubenswrapper[4696]: [+]process-running ok Mar 21 08:30:05 crc kubenswrapper[4696]: healthz check failed Mar 21 08:30:05 crc kubenswrapper[4696]: I0321 08:30:05.701727 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sdgwl" podUID="231fd5a3-3560-4a89-a5be-59880089a8fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 08:30:06 crc kubenswrapper[4696]: I0321 08:30:06.701016 4696 patch_prober.go:28] interesting pod/router-default-5444994796-sdgwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 08:30:06 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Mar 21 08:30:06 crc kubenswrapper[4696]: [+]process-running ok Mar 21 08:30:06 crc kubenswrapper[4696]: healthz check failed Mar 21 08:30:06 crc kubenswrapper[4696]: I0321 08:30:06.701104 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sdgwl" podUID="231fd5a3-3560-4a89-a5be-59880089a8fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 08:30:06 crc kubenswrapper[4696]: I0321 08:30:06.972410 4696 generic.go:334] "Generic (PLEG): container finished" podID="3fc5cd50-d5a4-4adf-9ef1-7894814b08c3" containerID="5d3c3d1af92fdcea421c2040343f9f31465c7268a8b2752a6a3923e1fff45744" exitCode=0 Mar 21 08:30:06 crc kubenswrapper[4696]: I0321 08:30:06.972483 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" event={"ID":"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3","Type":"ContainerDied","Data":"5d3c3d1af92fdcea421c2040343f9f31465c7268a8b2752a6a3923e1fff45744"} Mar 21 08:30:06 crc kubenswrapper[4696]: I0321 08:30:06.975105 4696 generic.go:334] "Generic (PLEG): container finished" podID="e1690a33-2e30-4c37-ba60-5def18cc3867" containerID="a1e9dd16fb92dbc127e8a88b556a14e21572f31a679145c5a16e034f058e914e" exitCode=0 Mar 21 08:30:06 crc kubenswrapper[4696]: I0321 08:30:06.975145 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" event={"ID":"e1690a33-2e30-4c37-ba60-5def18cc3867","Type":"ContainerDied","Data":"a1e9dd16fb92dbc127e8a88b556a14e21572f31a679145c5a16e034f058e914e"} Mar 21 08:30:07 crc kubenswrapper[4696]: I0321 08:30:07.318692 4696 ???:1] "http: TLS handshake error from 192.168.126.11:38302: no serving certificate available for the kubelet" Mar 21 08:30:07 crc kubenswrapper[4696]: I0321 08:30:07.701053 4696 patch_prober.go:28] interesting pod/router-default-5444994796-sdgwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 08:30:07 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Mar 21 08:30:07 crc kubenswrapper[4696]: [+]process-running ok Mar 21 08:30:07 crc kubenswrapper[4696]: healthz check failed Mar 21 08:30:07 crc kubenswrapper[4696]: I0321 08:30:07.701118 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sdgwl" podUID="231fd5a3-3560-4a89-a5be-59880089a8fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 08:30:08 crc kubenswrapper[4696]: I0321 08:30:08.701555 4696 patch_prober.go:28] interesting pod/router-default-5444994796-sdgwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 08:30:08 crc kubenswrapper[4696]: [-]has-synced failed: reason withheld Mar 21 08:30:08 crc kubenswrapper[4696]: [+]process-running ok Mar 21 08:30:08 crc kubenswrapper[4696]: healthz check failed Mar 21 08:30:08 crc kubenswrapper[4696]: I0321 08:30:08.701616 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sdgwl" podUID="231fd5a3-3560-4a89-a5be-59880089a8fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 08:30:09 crc kubenswrapper[4696]: I0321 08:30:09.701762 4696 patch_prober.go:28] interesting pod/router-default-5444994796-sdgwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 08:30:09 crc kubenswrapper[4696]: [+]has-synced ok Mar 21 08:30:09 crc kubenswrapper[4696]: [+]process-running ok Mar 21 08:30:09 crc kubenswrapper[4696]: healthz check failed Mar 21 08:30:09 crc kubenswrapper[4696]: I0321 08:30:09.701863 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sdgwl" podUID="231fd5a3-3560-4a89-a5be-59880089a8fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 08:30:10 crc kubenswrapper[4696]: I0321 08:30:10.704853 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-sdgwl" Mar 21 08:30:10 crc kubenswrapper[4696]: I0321 08:30:10.709349 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-sdgwl" Mar 21 08:30:10 crc kubenswrapper[4696]: I0321 08:30:10.937559 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 21 08:30:11 crc kubenswrapper[4696]: I0321 08:30:11.001579 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c14af961-f06f-4759-aff6-2b69ec8914e0","Type":"ContainerDied","Data":"a7a75348e92cefc3fb6fa6439ebf1aefe4f7d76bd1f3365fa335edeeb72d94cd"} Mar 21 08:30:11 crc kubenswrapper[4696]: I0321 08:30:11.001615 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 21 08:30:11 crc kubenswrapper[4696]: I0321 08:30:11.001620 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7a75348e92cefc3fb6fa6439ebf1aefe4f7d76bd1f3365fa335edeeb72d94cd" Mar 21 08:30:11 crc kubenswrapper[4696]: I0321 08:30:11.027079 4696 patch_prober.go:28] interesting pod/route-controller-manager-847bc8884c-559kj container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.48:8443/healthz\": dial tcp 10.217.0.48:8443: connect: connection refused" start-of-body= Mar 21 08:30:11 crc kubenswrapper[4696]: I0321 08:30:11.027157 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" podUID="e1690a33-2e30-4c37-ba60-5def18cc3867" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.48:8443/healthz\": dial tcp 10.217.0.48:8443: connect: connection refused" Mar 21 08:30:11 crc kubenswrapper[4696]: I0321 08:30:11.135381 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c14af961-f06f-4759-aff6-2b69ec8914e0-kube-api-access\") pod \"c14af961-f06f-4759-aff6-2b69ec8914e0\" (UID: \"c14af961-f06f-4759-aff6-2b69ec8914e0\") " Mar 21 08:30:11 crc kubenswrapper[4696]: I0321 08:30:11.135458 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c14af961-f06f-4759-aff6-2b69ec8914e0-kubelet-dir\") pod \"c14af961-f06f-4759-aff6-2b69ec8914e0\" (UID: \"c14af961-f06f-4759-aff6-2b69ec8914e0\") " Mar 21 08:30:11 crc kubenswrapper[4696]: I0321 08:30:11.135831 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c14af961-f06f-4759-aff6-2b69ec8914e0-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c14af961-f06f-4759-aff6-2b69ec8914e0" (UID: "c14af961-f06f-4759-aff6-2b69ec8914e0"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:30:11 crc kubenswrapper[4696]: I0321 08:30:11.141596 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c14af961-f06f-4759-aff6-2b69ec8914e0-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c14af961-f06f-4759-aff6-2b69ec8914e0" (UID: "c14af961-f06f-4759-aff6-2b69ec8914e0"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:30:11 crc kubenswrapper[4696]: I0321 08:30:11.237120 4696 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c14af961-f06f-4759-aff6-2b69ec8914e0-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:11 crc kubenswrapper[4696]: I0321 08:30:11.237149 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c14af961-f06f-4759-aff6-2b69ec8914e0-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:13 crc kubenswrapper[4696]: I0321 08:30:13.226407 4696 patch_prober.go:28] interesting pod/controller-manager-6c7d9ccbdf-9nbwj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.53:8443/healthz\": dial tcp 10.217.0.53:8443: connect: connection refused" start-of-body= Mar 21 08:30:13 crc kubenswrapper[4696]: I0321 08:30:13.226875 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" podUID="3fc5cd50-d5a4-4adf-9ef1-7894814b08c3" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.53:8443/healthz\": dial tcp 10.217.0.53:8443: connect: connection refused" Mar 21 08:30:13 crc kubenswrapper[4696]: I0321 08:30:13.369975 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:30:13 crc kubenswrapper[4696]: I0321 08:30:13.384601 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:30:13 crc kubenswrapper[4696]: I0321 08:30:13.651280 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568030-n2gf7" Mar 21 08:30:13 crc kubenswrapper[4696]: I0321 08:30:13.770581 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df64449e-f8aa-4e41-9b54-1dcc659d995c-secret-volume\") pod \"df64449e-f8aa-4e41-9b54-1dcc659d995c\" (UID: \"df64449e-f8aa-4e41-9b54-1dcc659d995c\") " Mar 21 08:30:13 crc kubenswrapper[4696]: I0321 08:30:13.770664 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prwvn\" (UniqueName: \"kubernetes.io/projected/df64449e-f8aa-4e41-9b54-1dcc659d995c-kube-api-access-prwvn\") pod \"df64449e-f8aa-4e41-9b54-1dcc659d995c\" (UID: \"df64449e-f8aa-4e41-9b54-1dcc659d995c\") " Mar 21 08:30:13 crc kubenswrapper[4696]: I0321 08:30:13.770801 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df64449e-f8aa-4e41-9b54-1dcc659d995c-config-volume\") pod \"df64449e-f8aa-4e41-9b54-1dcc659d995c\" (UID: \"df64449e-f8aa-4e41-9b54-1dcc659d995c\") " Mar 21 08:30:13 crc kubenswrapper[4696]: I0321 08:30:13.772089 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df64449e-f8aa-4e41-9b54-1dcc659d995c-config-volume" (OuterVolumeSpecName: "config-volume") pod "df64449e-f8aa-4e41-9b54-1dcc659d995c" (UID: "df64449e-f8aa-4e41-9b54-1dcc659d995c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:30:13 crc kubenswrapper[4696]: I0321 08:30:13.775352 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df64449e-f8aa-4e41-9b54-1dcc659d995c-kube-api-access-prwvn" (OuterVolumeSpecName: "kube-api-access-prwvn") pod "df64449e-f8aa-4e41-9b54-1dcc659d995c" (UID: "df64449e-f8aa-4e41-9b54-1dcc659d995c"). InnerVolumeSpecName "kube-api-access-prwvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:30:13 crc kubenswrapper[4696]: I0321 08:30:13.778340 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df64449e-f8aa-4e41-9b54-1dcc659d995c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "df64449e-f8aa-4e41-9b54-1dcc659d995c" (UID: "df64449e-f8aa-4e41-9b54-1dcc659d995c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:30:13 crc kubenswrapper[4696]: I0321 08:30:13.873148 4696 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df64449e-f8aa-4e41-9b54-1dcc659d995c-config-volume\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:13 crc kubenswrapper[4696]: I0321 08:30:13.873178 4696 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df64449e-f8aa-4e41-9b54-1dcc659d995c-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:13 crc kubenswrapper[4696]: I0321 08:30:13.873188 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prwvn\" (UniqueName: \"kubernetes.io/projected/df64449e-f8aa-4e41-9b54-1dcc659d995c-kube-api-access-prwvn\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:14 crc kubenswrapper[4696]: I0321 08:30:14.019188 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568030-n2gf7" Mar 21 08:30:14 crc kubenswrapper[4696]: I0321 08:30:14.027084 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568030-n2gf7" event={"ID":"df64449e-f8aa-4e41-9b54-1dcc659d995c","Type":"ContainerDied","Data":"2adf27cb8ae27e4af384395ee676b50a2d50b0539e4df3a753f1048e2fee98ed"} Mar 21 08:30:14 crc kubenswrapper[4696]: I0321 08:30:14.027160 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2adf27cb8ae27e4af384395ee676b50a2d50b0539e4df3a753f1048e2fee98ed" Mar 21 08:30:14 crc kubenswrapper[4696]: I0321 08:30:14.934161 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-jdfkn" Mar 21 08:30:15 crc kubenswrapper[4696]: E0321 08:30:15.124236 4696 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="430d7827b29f59b58867e896600db7736b683fcaad9bbd489d978811f63d0ec9" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 21 08:30:15 crc kubenswrapper[4696]: E0321 08:30:15.126101 4696 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="430d7827b29f59b58867e896600db7736b683fcaad9bbd489d978811f63d0ec9" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 21 08:30:15 crc kubenswrapper[4696]: E0321 08:30:15.127389 4696 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="430d7827b29f59b58867e896600db7736b683fcaad9bbd489d978811f63d0ec9" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 21 08:30:15 crc kubenswrapper[4696]: E0321 08:30:15.127423 4696 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" podUID="5f37526d-ea7c-4607-920a-ca63aa96184c" containerName="kube-multus-additional-cni-plugins" Mar 21 08:30:15 crc kubenswrapper[4696]: I0321 08:30:15.162295 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:30:22 crc kubenswrapper[4696]: I0321 08:30:22.027667 4696 patch_prober.go:28] interesting pod/route-controller-manager-847bc8884c-559kj container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.48:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 21 08:30:22 crc kubenswrapper[4696]: I0321 08:30:22.028776 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" podUID="e1690a33-2e30-4c37-ba60-5def18cc3867" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.48:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 21 08:30:23 crc kubenswrapper[4696]: I0321 08:30:23.510566 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 21 08:30:24 crc kubenswrapper[4696]: I0321 08:30:24.225650 4696 patch_prober.go:28] interesting pod/controller-manager-6c7d9ccbdf-9nbwj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.53:8443/healthz\": dial tcp 10.217.0.53:8443: i/o timeout" start-of-body= Mar 21 08:30:24 crc kubenswrapper[4696]: I0321 08:30:24.225729 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" podUID="3fc5cd50-d5a4-4adf-9ef1-7894814b08c3" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.53:8443/healthz\": dial tcp 10.217.0.53:8443: i/o timeout" Mar 21 08:30:24 crc kubenswrapper[4696]: I0321 08:30:24.796337 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qm5pb" Mar 21 08:30:25 crc kubenswrapper[4696]: I0321 08:30:25.080318 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-wkbtr_5f37526d-ea7c-4607-920a-ca63aa96184c/kube-multus-additional-cni-plugins/0.log" Mar 21 08:30:25 crc kubenswrapper[4696]: I0321 08:30:25.080557 4696 generic.go:334] "Generic (PLEG): container finished" podID="5f37526d-ea7c-4607-920a-ca63aa96184c" containerID="430d7827b29f59b58867e896600db7736b683fcaad9bbd489d978811f63d0ec9" exitCode=137 Mar 21 08:30:25 crc kubenswrapper[4696]: I0321 08:30:25.080594 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" event={"ID":"5f37526d-ea7c-4607-920a-ca63aa96184c","Type":"ContainerDied","Data":"430d7827b29f59b58867e896600db7736b683fcaad9bbd489d978811f63d0ec9"} Mar 21 08:30:25 crc kubenswrapper[4696]: E0321 08:30:25.121697 4696 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 430d7827b29f59b58867e896600db7736b683fcaad9bbd489d978811f63d0ec9 is running failed: container process not found" containerID="430d7827b29f59b58867e896600db7736b683fcaad9bbd489d978811f63d0ec9" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 21 08:30:25 crc kubenswrapper[4696]: E0321 08:30:25.122054 4696 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 430d7827b29f59b58867e896600db7736b683fcaad9bbd489d978811f63d0ec9 is running failed: container process not found" containerID="430d7827b29f59b58867e896600db7736b683fcaad9bbd489d978811f63d0ec9" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 21 08:30:25 crc kubenswrapper[4696]: E0321 08:30:25.122482 4696 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 430d7827b29f59b58867e896600db7736b683fcaad9bbd489d978811f63d0ec9 is running failed: container process not found" containerID="430d7827b29f59b58867e896600db7736b683fcaad9bbd489d978811f63d0ec9" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 21 08:30:25 crc kubenswrapper[4696]: E0321 08:30:25.122553 4696 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 430d7827b29f59b58867e896600db7736b683fcaad9bbd489d978811f63d0ec9 is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" podUID="5f37526d-ea7c-4607-920a-ca63aa96184c" containerName="kube-multus-additional-cni-plugins" Mar 21 08:30:26 crc kubenswrapper[4696]: I0321 08:30:26.557197 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 21 08:30:27 crc kubenswrapper[4696]: I0321 08:30:27.830147 4696 ???:1] "http: TLS handshake error from 192.168.126.11:50656: no serving certificate available for the kubelet" Mar 21 08:30:27 crc kubenswrapper[4696]: E0321 08:30:27.832366 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 21 08:30:27 crc kubenswrapper[4696]: E0321 08:30:27.832541 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n825s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-cfjs8_openshift-marketplace(0030cc8a-f473-4d0a-ace9-728b66e62634): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 21 08:30:27 crc kubenswrapper[4696]: E0321 08:30:27.833659 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-cfjs8" podUID="0030cc8a-f473-4d0a-ace9-728b66e62634" Mar 21 08:30:28 crc kubenswrapper[4696]: I0321 08:30:28.156089 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=2.156063756 podStartE2EDuration="2.156063756s" podCreationTimestamp="2026-03-21 08:30:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:30:28.140352968 +0000 UTC m=+162.261233671" watchObservedRunningTime="2026-03-21 08:30:28.156063756 +0000 UTC m=+162.276944469" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.194162 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 21 08:30:29 crc kubenswrapper[4696]: E0321 08:30:29.194539 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6836f01-7f9f-4cc7-99d8-10eababb2b1f" containerName="pruner" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.194551 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6836f01-7f9f-4cc7-99d8-10eababb2b1f" containerName="pruner" Mar 21 08:30:29 crc kubenswrapper[4696]: E0321 08:30:29.194563 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c14af961-f06f-4759-aff6-2b69ec8914e0" containerName="pruner" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.194585 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="c14af961-f06f-4759-aff6-2b69ec8914e0" containerName="pruner" Mar 21 08:30:29 crc kubenswrapper[4696]: E0321 08:30:29.194604 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df64449e-f8aa-4e41-9b54-1dcc659d995c" containerName="collect-profiles" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.194610 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="df64449e-f8aa-4e41-9b54-1dcc659d995c" containerName="collect-profiles" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.194740 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="c14af961-f06f-4759-aff6-2b69ec8914e0" containerName="pruner" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.194751 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6836f01-7f9f-4cc7-99d8-10eababb2b1f" containerName="pruner" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.194761 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="df64449e-f8aa-4e41-9b54-1dcc659d995c" containerName="collect-profiles" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.195254 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.200449 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.200512 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.201006 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.302900 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/734c53f5-4c0a-4948-aa09-b8c20ce77920-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"734c53f5-4c0a-4948-aa09-b8c20ce77920\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.303322 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/734c53f5-4c0a-4948-aa09-b8c20ce77920-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"734c53f5-4c0a-4948-aa09-b8c20ce77920\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.404848 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/734c53f5-4c0a-4948-aa09-b8c20ce77920-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"734c53f5-4c0a-4948-aa09-b8c20ce77920\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.404910 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/734c53f5-4c0a-4948-aa09-b8c20ce77920-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"734c53f5-4c0a-4948-aa09-b8c20ce77920\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.405037 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/734c53f5-4c0a-4948-aa09-b8c20ce77920-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"734c53f5-4c0a-4948-aa09-b8c20ce77920\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.422596 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/734c53f5-4c0a-4948-aa09-b8c20ce77920-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"734c53f5-4c0a-4948-aa09-b8c20ce77920\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.521886 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 21 08:30:29 crc kubenswrapper[4696]: E0321 08:30:29.591623 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-cfjs8" podUID="0030cc8a-f473-4d0a-ace9-728b66e62634" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.640509 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.653101 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" Mar 21 08:30:29 crc kubenswrapper[4696]: E0321 08:30:29.665792 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 21 08:30:29 crc kubenswrapper[4696]: E0321 08:30:29.666003 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7gs9s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-7tpgm_openshift-marketplace(5149cc5e-d877-4d52-8db6-7df6ae060ab0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 21 08:30:29 crc kubenswrapper[4696]: E0321 08:30:29.667172 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-7tpgm" podUID="5149cc5e-d877-4d52-8db6-7df6ae060ab0" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.685331 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6"] Mar 21 08:30:29 crc kubenswrapper[4696]: E0321 08:30:29.685655 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fc5cd50-d5a4-4adf-9ef1-7894814b08c3" containerName="controller-manager" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.685669 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fc5cd50-d5a4-4adf-9ef1-7894814b08c3" containerName="controller-manager" Mar 21 08:30:29 crc kubenswrapper[4696]: E0321 08:30:29.685694 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1690a33-2e30-4c37-ba60-5def18cc3867" containerName="route-controller-manager" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.685702 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1690a33-2e30-4c37-ba60-5def18cc3867" containerName="route-controller-manager" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.685833 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1690a33-2e30-4c37-ba60-5def18cc3867" containerName="route-controller-manager" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.685852 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fc5cd50-d5a4-4adf-9ef1-7894814b08c3" containerName="controller-manager" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.686354 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.689451 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6"] Mar 21 08:30:29 crc kubenswrapper[4696]: E0321 08:30:29.694484 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 21 08:30:29 crc kubenswrapper[4696]: E0321 08:30:29.694616 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dpfnx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-trvd2_openshift-marketplace(d6b6d345-41c3-4926-bf33-bac67ac54755): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 21 08:30:29 crc kubenswrapper[4696]: E0321 08:30:29.696846 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-trvd2" podUID="d6b6d345-41c3-4926-bf33-bac67ac54755" Mar 21 08:30:29 crc kubenswrapper[4696]: E0321 08:30:29.743949 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 21 08:30:29 crc kubenswrapper[4696]: E0321 08:30:29.744111 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9q4bs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-jmqr7_openshift-marketplace(6ed31bad-6fb5-49be-922b-27198867ca4a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 21 08:30:29 crc kubenswrapper[4696]: E0321 08:30:29.745259 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-jmqr7" podUID="6ed31bad-6fb5-49be-922b-27198867ca4a" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.810071 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e1690a33-2e30-4c37-ba60-5def18cc3867-client-ca\") pod \"e1690a33-2e30-4c37-ba60-5def18cc3867\" (UID: \"e1690a33-2e30-4c37-ba60-5def18cc3867\") " Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.810120 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-serving-cert\") pod \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\" (UID: \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\") " Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.810179 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-proxy-ca-bundles\") pod \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\" (UID: \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\") " Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.810254 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wh7kh\" (UniqueName: \"kubernetes.io/projected/e1690a33-2e30-4c37-ba60-5def18cc3867-kube-api-access-wh7kh\") pod \"e1690a33-2e30-4c37-ba60-5def18cc3867\" (UID: \"e1690a33-2e30-4c37-ba60-5def18cc3867\") " Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.810302 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvchl\" (UniqueName: \"kubernetes.io/projected/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-kube-api-access-rvchl\") pod \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\" (UID: \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\") " Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.810338 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1690a33-2e30-4c37-ba60-5def18cc3867-config\") pod \"e1690a33-2e30-4c37-ba60-5def18cc3867\" (UID: \"e1690a33-2e30-4c37-ba60-5def18cc3867\") " Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.810399 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-config\") pod \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\" (UID: \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\") " Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.810424 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-client-ca\") pod \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\" (UID: \"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3\") " Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.810456 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1690a33-2e30-4c37-ba60-5def18cc3867-serving-cert\") pod \"e1690a33-2e30-4c37-ba60-5def18cc3867\" (UID: \"e1690a33-2e30-4c37-ba60-5def18cc3867\") " Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.810663 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e77a6440-cb75-4a20-bc27-ee519003841b-config\") pod \"controller-manager-7dbddf9fd5-td2n6\" (UID: \"e77a6440-cb75-4a20-bc27-ee519003841b\") " pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.810702 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e77a6440-cb75-4a20-bc27-ee519003841b-proxy-ca-bundles\") pod \"controller-manager-7dbddf9fd5-td2n6\" (UID: \"e77a6440-cb75-4a20-bc27-ee519003841b\") " pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.810730 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e77a6440-cb75-4a20-bc27-ee519003841b-serving-cert\") pod \"controller-manager-7dbddf9fd5-td2n6\" (UID: \"e77a6440-cb75-4a20-bc27-ee519003841b\") " pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.810750 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrwvf\" (UniqueName: \"kubernetes.io/projected/e77a6440-cb75-4a20-bc27-ee519003841b-kube-api-access-wrwvf\") pod \"controller-manager-7dbddf9fd5-td2n6\" (UID: \"e77a6440-cb75-4a20-bc27-ee519003841b\") " pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.810775 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e77a6440-cb75-4a20-bc27-ee519003841b-client-ca\") pod \"controller-manager-7dbddf9fd5-td2n6\" (UID: \"e77a6440-cb75-4a20-bc27-ee519003841b\") " pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.811535 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-client-ca" (OuterVolumeSpecName: "client-ca") pod "3fc5cd50-d5a4-4adf-9ef1-7894814b08c3" (UID: "3fc5cd50-d5a4-4adf-9ef1-7894814b08c3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.811571 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-config" (OuterVolumeSpecName: "config") pod "3fc5cd50-d5a4-4adf-9ef1-7894814b08c3" (UID: "3fc5cd50-d5a4-4adf-9ef1-7894814b08c3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.811694 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1690a33-2e30-4c37-ba60-5def18cc3867-client-ca" (OuterVolumeSpecName: "client-ca") pod "e1690a33-2e30-4c37-ba60-5def18cc3867" (UID: "e1690a33-2e30-4c37-ba60-5def18cc3867"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.812538 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1690a33-2e30-4c37-ba60-5def18cc3867-config" (OuterVolumeSpecName: "config") pod "e1690a33-2e30-4c37-ba60-5def18cc3867" (UID: "e1690a33-2e30-4c37-ba60-5def18cc3867"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.817200 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-kube-api-access-rvchl" (OuterVolumeSpecName: "kube-api-access-rvchl") pod "3fc5cd50-d5a4-4adf-9ef1-7894814b08c3" (UID: "3fc5cd50-d5a4-4adf-9ef1-7894814b08c3"). InnerVolumeSpecName "kube-api-access-rvchl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.827320 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1690a33-2e30-4c37-ba60-5def18cc3867-kube-api-access-wh7kh" (OuterVolumeSpecName: "kube-api-access-wh7kh") pod "e1690a33-2e30-4c37-ba60-5def18cc3867" (UID: "e1690a33-2e30-4c37-ba60-5def18cc3867"). InnerVolumeSpecName "kube-api-access-wh7kh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.827389 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "3fc5cd50-d5a4-4adf-9ef1-7894814b08c3" (UID: "3fc5cd50-d5a4-4adf-9ef1-7894814b08c3"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.828833 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1690a33-2e30-4c37-ba60-5def18cc3867-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e1690a33-2e30-4c37-ba60-5def18cc3867" (UID: "e1690a33-2e30-4c37-ba60-5def18cc3867"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.833621 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3fc5cd50-d5a4-4adf-9ef1-7894814b08c3" (UID: "3fc5cd50-d5a4-4adf-9ef1-7894814b08c3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.912173 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e77a6440-cb75-4a20-bc27-ee519003841b-serving-cert\") pod \"controller-manager-7dbddf9fd5-td2n6\" (UID: \"e77a6440-cb75-4a20-bc27-ee519003841b\") " pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.912241 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrwvf\" (UniqueName: \"kubernetes.io/projected/e77a6440-cb75-4a20-bc27-ee519003841b-kube-api-access-wrwvf\") pod \"controller-manager-7dbddf9fd5-td2n6\" (UID: \"e77a6440-cb75-4a20-bc27-ee519003841b\") " pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.912273 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e77a6440-cb75-4a20-bc27-ee519003841b-client-ca\") pod \"controller-manager-7dbddf9fd5-td2n6\" (UID: \"e77a6440-cb75-4a20-bc27-ee519003841b\") " pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.912360 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e77a6440-cb75-4a20-bc27-ee519003841b-config\") pod \"controller-manager-7dbddf9fd5-td2n6\" (UID: \"e77a6440-cb75-4a20-bc27-ee519003841b\") " pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.912401 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e77a6440-cb75-4a20-bc27-ee519003841b-proxy-ca-bundles\") pod \"controller-manager-7dbddf9fd5-td2n6\" (UID: \"e77a6440-cb75-4a20-bc27-ee519003841b\") " pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.912457 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wh7kh\" (UniqueName: \"kubernetes.io/projected/e1690a33-2e30-4c37-ba60-5def18cc3867-kube-api-access-wh7kh\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.912583 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvchl\" (UniqueName: \"kubernetes.io/projected/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-kube-api-access-rvchl\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.913063 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1690a33-2e30-4c37-ba60-5def18cc3867-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.913233 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.913370 4696 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.913515 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1690a33-2e30-4c37-ba60-5def18cc3867-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.913651 4696 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e1690a33-2e30-4c37-ba60-5def18cc3867-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.913775 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.913682 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e77a6440-cb75-4a20-bc27-ee519003841b-config\") pod \"controller-manager-7dbddf9fd5-td2n6\" (UID: \"e77a6440-cb75-4a20-bc27-ee519003841b\") " pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.913933 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e77a6440-cb75-4a20-bc27-ee519003841b-proxy-ca-bundles\") pod \"controller-manager-7dbddf9fd5-td2n6\" (UID: \"e77a6440-cb75-4a20-bc27-ee519003841b\") " pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.913386 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e77a6440-cb75-4a20-bc27-ee519003841b-client-ca\") pod \"controller-manager-7dbddf9fd5-td2n6\" (UID: \"e77a6440-cb75-4a20-bc27-ee519003841b\") " pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.913937 4696 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.915263 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e77a6440-cb75-4a20-bc27-ee519003841b-serving-cert\") pod \"controller-manager-7dbddf9fd5-td2n6\" (UID: \"e77a6440-cb75-4a20-bc27-ee519003841b\") " pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" Mar 21 08:30:29 crc kubenswrapper[4696]: I0321 08:30:29.930227 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrwvf\" (UniqueName: \"kubernetes.io/projected/e77a6440-cb75-4a20-bc27-ee519003841b-kube-api-access-wrwvf\") pod \"controller-manager-7dbddf9fd5-td2n6\" (UID: \"e77a6440-cb75-4a20-bc27-ee519003841b\") " pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" Mar 21 08:30:30 crc kubenswrapper[4696]: I0321 08:30:30.007238 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" Mar 21 08:30:30 crc kubenswrapper[4696]: I0321 08:30:30.111268 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" Mar 21 08:30:30 crc kubenswrapper[4696]: I0321 08:30:30.111260 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj" event={"ID":"e1690a33-2e30-4c37-ba60-5def18cc3867","Type":"ContainerDied","Data":"4d3336bfb1e29e14fab43b72b48a3ca45d88fe9bf1e608647f15c5020407072a"} Mar 21 08:30:30 crc kubenswrapper[4696]: I0321 08:30:30.112280 4696 scope.go:117] "RemoveContainer" containerID="a1e9dd16fb92dbc127e8a88b556a14e21572f31a679145c5a16e034f058e914e" Mar 21 08:30:30 crc kubenswrapper[4696]: I0321 08:30:30.115121 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" Mar 21 08:30:30 crc kubenswrapper[4696]: I0321 08:30:30.117273 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj" event={"ID":"3fc5cd50-d5a4-4adf-9ef1-7894814b08c3","Type":"ContainerDied","Data":"ce27886d958a66c78ea7409517fd0e24938be118ca5d08a776a5533031dbe1b9"} Mar 21 08:30:30 crc kubenswrapper[4696]: I0321 08:30:30.193118 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj"] Mar 21 08:30:30 crc kubenswrapper[4696]: I0321 08:30:30.200322 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6c7d9ccbdf-9nbwj"] Mar 21 08:30:30 crc kubenswrapper[4696]: I0321 08:30:30.204361 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj"] Mar 21 08:30:30 crc kubenswrapper[4696]: I0321 08:30:30.208573 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-847bc8884c-559kj"] Mar 21 08:30:30 crc kubenswrapper[4696]: I0321 08:30:30.542588 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fc5cd50-d5a4-4adf-9ef1-7894814b08c3" path="/var/lib/kubelet/pods/3fc5cd50-d5a4-4adf-9ef1-7894814b08c3/volumes" Mar 21 08:30:30 crc kubenswrapper[4696]: I0321 08:30:30.543163 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1690a33-2e30-4c37-ba60-5def18cc3867" path="/var/lib/kubelet/pods/e1690a33-2e30-4c37-ba60-5def18cc3867/volumes" Mar 21 08:30:31 crc kubenswrapper[4696]: I0321 08:30:31.916359 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt"] Mar 21 08:30:31 crc kubenswrapper[4696]: I0321 08:30:31.917656 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" Mar 21 08:30:31 crc kubenswrapper[4696]: I0321 08:30:31.919538 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 21 08:30:31 crc kubenswrapper[4696]: I0321 08:30:31.919827 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 21 08:30:31 crc kubenswrapper[4696]: I0321 08:30:31.920065 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 21 08:30:31 crc kubenswrapper[4696]: I0321 08:30:31.920144 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 21 08:30:31 crc kubenswrapper[4696]: I0321 08:30:31.920338 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 21 08:30:31 crc kubenswrapper[4696]: I0321 08:30:31.920552 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 21 08:30:31 crc kubenswrapper[4696]: E0321 08:30:31.963108 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-7tpgm" podUID="5149cc5e-d877-4d52-8db6-7df6ae060ab0" Mar 21 08:30:31 crc kubenswrapper[4696]: E0321 08:30:31.963634 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-trvd2" podUID="d6b6d345-41c3-4926-bf33-bac67ac54755" Mar 21 08:30:31 crc kubenswrapper[4696]: I0321 08:30:31.964196 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ntrb\" (UniqueName: \"kubernetes.io/projected/d6fa2cee-b7db-4516-997f-1602ecf67afb-kube-api-access-5ntrb\") pod \"route-controller-manager-74dfdbbfc5-kxrnt\" (UID: \"d6fa2cee-b7db-4516-997f-1602ecf67afb\") " pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" Mar 21 08:30:31 crc kubenswrapper[4696]: I0321 08:30:31.964241 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d6fa2cee-b7db-4516-997f-1602ecf67afb-client-ca\") pod \"route-controller-manager-74dfdbbfc5-kxrnt\" (UID: \"d6fa2cee-b7db-4516-997f-1602ecf67afb\") " pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" Mar 21 08:30:31 crc kubenswrapper[4696]: I0321 08:30:31.964303 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6fa2cee-b7db-4516-997f-1602ecf67afb-serving-cert\") pod \"route-controller-manager-74dfdbbfc5-kxrnt\" (UID: \"d6fa2cee-b7db-4516-997f-1602ecf67afb\") " pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" Mar 21 08:30:31 crc kubenswrapper[4696]: E0321 08:30:31.964321 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-jmqr7" podUID="6ed31bad-6fb5-49be-922b-27198867ca4a" Mar 21 08:30:31 crc kubenswrapper[4696]: I0321 08:30:31.964380 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6fa2cee-b7db-4516-997f-1602ecf67afb-config\") pod \"route-controller-manager-74dfdbbfc5-kxrnt\" (UID: \"d6fa2cee-b7db-4516-997f-1602ecf67afb\") " pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" Mar 21 08:30:31 crc kubenswrapper[4696]: I0321 08:30:31.968064 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt"] Mar 21 08:30:32 crc kubenswrapper[4696]: E0321 08:30:32.062908 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 21 08:30:32 crc kubenswrapper[4696]: E0321 08:30:32.063387 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gswrs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-qsg67_openshift-marketplace(6aabaca2-7def-4550-bae1-9337b7c1002b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 21 08:30:32 crc kubenswrapper[4696]: E0321 08:30:32.064558 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-qsg67" podUID="6aabaca2-7def-4550-bae1-9337b7c1002b" Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.065122 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6fa2cee-b7db-4516-997f-1602ecf67afb-config\") pod \"route-controller-manager-74dfdbbfc5-kxrnt\" (UID: \"d6fa2cee-b7db-4516-997f-1602ecf67afb\") " pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.065181 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ntrb\" (UniqueName: \"kubernetes.io/projected/d6fa2cee-b7db-4516-997f-1602ecf67afb-kube-api-access-5ntrb\") pod \"route-controller-manager-74dfdbbfc5-kxrnt\" (UID: \"d6fa2cee-b7db-4516-997f-1602ecf67afb\") " pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.065210 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d6fa2cee-b7db-4516-997f-1602ecf67afb-client-ca\") pod \"route-controller-manager-74dfdbbfc5-kxrnt\" (UID: \"d6fa2cee-b7db-4516-997f-1602ecf67afb\") " pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.065248 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6fa2cee-b7db-4516-997f-1602ecf67afb-serving-cert\") pod \"route-controller-manager-74dfdbbfc5-kxrnt\" (UID: \"d6fa2cee-b7db-4516-997f-1602ecf67afb\") " pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.067768 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6fa2cee-b7db-4516-997f-1602ecf67afb-config\") pod \"route-controller-manager-74dfdbbfc5-kxrnt\" (UID: \"d6fa2cee-b7db-4516-997f-1602ecf67afb\") " pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.068183 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d6fa2cee-b7db-4516-997f-1602ecf67afb-client-ca\") pod \"route-controller-manager-74dfdbbfc5-kxrnt\" (UID: \"d6fa2cee-b7db-4516-997f-1602ecf67afb\") " pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.081987 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ntrb\" (UniqueName: \"kubernetes.io/projected/d6fa2cee-b7db-4516-997f-1602ecf67afb-kube-api-access-5ntrb\") pod \"route-controller-manager-74dfdbbfc5-kxrnt\" (UID: \"d6fa2cee-b7db-4516-997f-1602ecf67afb\") " pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.083495 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6fa2cee-b7db-4516-997f-1602ecf67afb-serving-cert\") pod \"route-controller-manager-74dfdbbfc5-kxrnt\" (UID: \"d6fa2cee-b7db-4516-997f-1602ecf67afb\") " pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.295199 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" Mar 21 08:30:32 crc kubenswrapper[4696]: E0321 08:30:32.358140 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-qsg67" podUID="6aabaca2-7def-4550-bae1-9337b7c1002b" Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.372968 4696 scope.go:117] "RemoveContainer" containerID="5d3c3d1af92fdcea421c2040343f9f31465c7268a8b2752a6a3923e1fff45744" Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.414601 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-wkbtr_5f37526d-ea7c-4607-920a-ca63aa96184c/kube-multus-additional-cni-plugins/0.log" Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.414677 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.573457 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5f37526d-ea7c-4607-920a-ca63aa96184c-cni-sysctl-allowlist\") pod \"5f37526d-ea7c-4607-920a-ca63aa96184c\" (UID: \"5f37526d-ea7c-4607-920a-ca63aa96184c\") " Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.573597 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5f37526d-ea7c-4607-920a-ca63aa96184c-tuning-conf-dir\") pod \"5f37526d-ea7c-4607-920a-ca63aa96184c\" (UID: \"5f37526d-ea7c-4607-920a-ca63aa96184c\") " Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.573635 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcpl7\" (UniqueName: \"kubernetes.io/projected/5f37526d-ea7c-4607-920a-ca63aa96184c-kube-api-access-vcpl7\") pod \"5f37526d-ea7c-4607-920a-ca63aa96184c\" (UID: \"5f37526d-ea7c-4607-920a-ca63aa96184c\") " Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.573667 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/5f37526d-ea7c-4607-920a-ca63aa96184c-ready\") pod \"5f37526d-ea7c-4607-920a-ca63aa96184c\" (UID: \"5f37526d-ea7c-4607-920a-ca63aa96184c\") " Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.574354 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f37526d-ea7c-4607-920a-ca63aa96184c-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "5f37526d-ea7c-4607-920a-ca63aa96184c" (UID: "5f37526d-ea7c-4607-920a-ca63aa96184c"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.574919 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f37526d-ea7c-4607-920a-ca63aa96184c-ready" (OuterVolumeSpecName: "ready") pod "5f37526d-ea7c-4607-920a-ca63aa96184c" (UID: "5f37526d-ea7c-4607-920a-ca63aa96184c"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.574945 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f37526d-ea7c-4607-920a-ca63aa96184c-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "5f37526d-ea7c-4607-920a-ca63aa96184c" (UID: "5f37526d-ea7c-4607-920a-ca63aa96184c"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.578431 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f37526d-ea7c-4607-920a-ca63aa96184c-kube-api-access-vcpl7" (OuterVolumeSpecName: "kube-api-access-vcpl7") pod "5f37526d-ea7c-4607-920a-ca63aa96184c" (UID: "5f37526d-ea7c-4607-920a-ca63aa96184c"). InnerVolumeSpecName "kube-api-access-vcpl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.675625 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcpl7\" (UniqueName: \"kubernetes.io/projected/5f37526d-ea7c-4607-920a-ca63aa96184c-kube-api-access-vcpl7\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.675666 4696 reconciler_common.go:293] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/5f37526d-ea7c-4607-920a-ca63aa96184c-ready\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.675680 4696 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5f37526d-ea7c-4607-920a-ca63aa96184c-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:32 crc kubenswrapper[4696]: I0321 08:30:32.675691 4696 reconciler_common.go:293] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5f37526d-ea7c-4607-920a-ca63aa96184c-tuning-conf-dir\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.132212 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-wkbtr_5f37526d-ea7c-4607-920a-ca63aa96184c/kube-multus-additional-cni-plugins/0.log" Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.132426 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" event={"ID":"5f37526d-ea7c-4607-920a-ca63aa96184c","Type":"ContainerDied","Data":"63c099de631eb958ac858b63b909c7a4ea802abf54198d5076f1512239135162"} Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.132493 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-wkbtr" Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.158749 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-wkbtr"] Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.165338 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-wkbtr"] Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.294002 4696 scope.go:117] "RemoveContainer" containerID="430d7827b29f59b58867e896600db7736b683fcaad9bbd489d978811f63d0ec9" Mar 21 08:30:33 crc kubenswrapper[4696]: E0321 08:30:33.327503 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Mar 21 08:30:33 crc kubenswrapper[4696]: E0321 08:30:33.327685 4696 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 21 08:30:33 crc kubenswrapper[4696]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Mar 21 08:30:33 crc kubenswrapper[4696]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bgqbs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29568030-4x55c_openshift-infra(a6581056-7677-491f-8f34-5626b43b6e77): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Mar 21 08:30:33 crc kubenswrapper[4696]: > logger="UnhandledError" Mar 21 08:30:33 crc kubenswrapper[4696]: E0321 08:30:33.328851 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29568030-4x55c" podUID="a6581056-7677-491f-8f34-5626b43b6e77" Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.567949 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6"] Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.585314 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 21 08:30:33 crc kubenswrapper[4696]: E0321 08:30:33.585626 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f37526d-ea7c-4607-920a-ca63aa96184c" containerName="kube-multus-additional-cni-plugins" Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.585640 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f37526d-ea7c-4607-920a-ca63aa96184c" containerName="kube-multus-additional-cni-plugins" Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.585739 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f37526d-ea7c-4607-920a-ca63aa96184c" containerName="kube-multus-additional-cni-plugins" Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.586132 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.600179 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.689223 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/df078e35-cb4d-4be7-beaf-7fd6fc463a0c-var-lock\") pod \"installer-9-crc\" (UID: \"df078e35-cb4d-4be7-beaf-7fd6fc463a0c\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.689276 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/df078e35-cb4d-4be7-beaf-7fd6fc463a0c-kube-api-access\") pod \"installer-9-crc\" (UID: \"df078e35-cb4d-4be7-beaf-7fd6fc463a0c\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.689747 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/df078e35-cb4d-4be7-beaf-7fd6fc463a0c-kubelet-dir\") pod \"installer-9-crc\" (UID: \"df078e35-cb4d-4be7-beaf-7fd6fc463a0c\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.723162 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.730215 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt"] Mar 21 08:30:33 crc kubenswrapper[4696]: W0321 08:30:33.773127 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6fa2cee_b7db_4516_997f_1602ecf67afb.slice/crio-2147be457b9cb2b43807422306b17df406016caa106ea12320ef643120523568 WatchSource:0}: Error finding container 2147be457b9cb2b43807422306b17df406016caa106ea12320ef643120523568: Status 404 returned error can't find the container with id 2147be457b9cb2b43807422306b17df406016caa106ea12320ef643120523568 Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.790966 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/df078e35-cb4d-4be7-beaf-7fd6fc463a0c-kubelet-dir\") pod \"installer-9-crc\" (UID: \"df078e35-cb4d-4be7-beaf-7fd6fc463a0c\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.791021 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/df078e35-cb4d-4be7-beaf-7fd6fc463a0c-var-lock\") pod \"installer-9-crc\" (UID: \"df078e35-cb4d-4be7-beaf-7fd6fc463a0c\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.791035 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/df078e35-cb4d-4be7-beaf-7fd6fc463a0c-kubelet-dir\") pod \"installer-9-crc\" (UID: \"df078e35-cb4d-4be7-beaf-7fd6fc463a0c\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.791056 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/df078e35-cb4d-4be7-beaf-7fd6fc463a0c-kube-api-access\") pod \"installer-9-crc\" (UID: \"df078e35-cb4d-4be7-beaf-7fd6fc463a0c\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.791168 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/df078e35-cb4d-4be7-beaf-7fd6fc463a0c-var-lock\") pod \"installer-9-crc\" (UID: \"df078e35-cb4d-4be7-beaf-7fd6fc463a0c\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.807692 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/df078e35-cb4d-4be7-beaf-7fd6fc463a0c-kube-api-access\") pod \"installer-9-crc\" (UID: \"df078e35-cb4d-4be7-beaf-7fd6fc463a0c\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 21 08:30:33 crc kubenswrapper[4696]: W0321 08:30:33.843944 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod734c53f5_4c0a_4948_aa09_b8c20ce77920.slice/crio-5157dddcd4907f20df02d18b5f7e04bac2f2901177dcfa6c800d6f0b732cec94 WatchSource:0}: Error finding container 5157dddcd4907f20df02d18b5f7e04bac2f2901177dcfa6c800d6f0b732cec94: Status 404 returned error can't find the container with id 5157dddcd4907f20df02d18b5f7e04bac2f2901177dcfa6c800d6f0b732cec94 Mar 21 08:30:33 crc kubenswrapper[4696]: I0321 08:30:33.901907 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 21 08:30:34 crc kubenswrapper[4696]: I0321 08:30:34.165594 4696 generic.go:334] "Generic (PLEG): container finished" podID="85a5000b-3b68-4808-bcfc-93bcbde10dde" containerID="d1d423895bc1e7afe3dda0ff1a30e13f1deafd6c471f6b26594f023d3828ec23" exitCode=0 Mar 21 08:30:34 crc kubenswrapper[4696]: I0321 08:30:34.165681 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v9rlg" event={"ID":"85a5000b-3b68-4808-bcfc-93bcbde10dde","Type":"ContainerDied","Data":"d1d423895bc1e7afe3dda0ff1a30e13f1deafd6c471f6b26594f023d3828ec23"} Mar 21 08:30:34 crc kubenswrapper[4696]: I0321 08:30:34.174189 4696 generic.go:334] "Generic (PLEG): container finished" podID="d70a46f0-1e5d-4edb-8745-9b8a41b66241" containerID="d29438d9bfb377184180255d89e71c9df999fdb0b903dc06d7601c2fc7d6a1a3" exitCode=0 Mar 21 08:30:34 crc kubenswrapper[4696]: I0321 08:30:34.174622 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jnw8m" event={"ID":"d70a46f0-1e5d-4edb-8745-9b8a41b66241","Type":"ContainerDied","Data":"d29438d9bfb377184180255d89e71c9df999fdb0b903dc06d7601c2fc7d6a1a3"} Mar 21 08:30:34 crc kubenswrapper[4696]: I0321 08:30:34.177076 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 21 08:30:34 crc kubenswrapper[4696]: I0321 08:30:34.178521 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"734c53f5-4c0a-4948-aa09-b8c20ce77920","Type":"ContainerStarted","Data":"5157dddcd4907f20df02d18b5f7e04bac2f2901177dcfa6c800d6f0b732cec94"} Mar 21 08:30:34 crc kubenswrapper[4696]: I0321 08:30:34.189162 4696 generic.go:334] "Generic (PLEG): container finished" podID="0b88fc0a-dc29-4aa2-b625-c24162c0dc0b" containerID="f8e612f0cdd24c6280d8c802539db297637f3450bc73471405923922ea03ed0d" exitCode=0 Mar 21 08:30:34 crc kubenswrapper[4696]: I0321 08:30:34.189257 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7wvkn" event={"ID":"0b88fc0a-dc29-4aa2-b625-c24162c0dc0b","Type":"ContainerDied","Data":"f8e612f0cdd24c6280d8c802539db297637f3450bc73471405923922ea03ed0d"} Mar 21 08:30:34 crc kubenswrapper[4696]: I0321 08:30:34.196520 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" event={"ID":"e77a6440-cb75-4a20-bc27-ee519003841b","Type":"ContainerStarted","Data":"1d832e1b53602f59f869850faadcbfdb0912be4ba51e640d93b76c75221d9122"} Mar 21 08:30:34 crc kubenswrapper[4696]: I0321 08:30:34.196595 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" event={"ID":"e77a6440-cb75-4a20-bc27-ee519003841b","Type":"ContainerStarted","Data":"2deafa2e79cddd590771b15bb45d6e740040bea32ae603d9ee3649c6dff2eb29"} Mar 21 08:30:34 crc kubenswrapper[4696]: I0321 08:30:34.196609 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" Mar 21 08:30:34 crc kubenswrapper[4696]: I0321 08:30:34.205032 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" event={"ID":"d6fa2cee-b7db-4516-997f-1602ecf67afb","Type":"ContainerStarted","Data":"1b3fe6812a90966c6ced0db1da6ee7f83691cb0d2661624483dc1ce8bae0eef1"} Mar 21 08:30:34 crc kubenswrapper[4696]: I0321 08:30:34.205069 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" event={"ID":"d6fa2cee-b7db-4516-997f-1602ecf67afb","Type":"ContainerStarted","Data":"2147be457b9cb2b43807422306b17df406016caa106ea12320ef643120523568"} Mar 21 08:30:34 crc kubenswrapper[4696]: I0321 08:30:34.205083 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" Mar 21 08:30:34 crc kubenswrapper[4696]: E0321 08:30:34.205655 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29568030-4x55c" podUID="a6581056-7677-491f-8f34-5626b43b6e77" Mar 21 08:30:34 crc kubenswrapper[4696]: I0321 08:30:34.207056 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" Mar 21 08:30:34 crc kubenswrapper[4696]: I0321 08:30:34.235099 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" podStartSLOduration=9.235078394 podStartE2EDuration="9.235078394s" podCreationTimestamp="2026-03-21 08:30:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:30:34.23380812 +0000 UTC m=+168.354688833" watchObservedRunningTime="2026-03-21 08:30:34.235078394 +0000 UTC m=+168.355959107" Mar 21 08:30:34 crc kubenswrapper[4696]: I0321 08:30:34.303553 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" podStartSLOduration=9.303533222 podStartE2EDuration="9.303533222s" podCreationTimestamp="2026-03-21 08:30:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:30:34.272291922 +0000 UTC m=+168.393172635" watchObservedRunningTime="2026-03-21 08:30:34.303533222 +0000 UTC m=+168.424413935" Mar 21 08:30:34 crc kubenswrapper[4696]: I0321 08:30:34.381429 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" Mar 21 08:30:34 crc kubenswrapper[4696]: I0321 08:30:34.546961 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f37526d-ea7c-4607-920a-ca63aa96184c" path="/var/lib/kubelet/pods/5f37526d-ea7c-4607-920a-ca63aa96184c/volumes" Mar 21 08:30:34 crc kubenswrapper[4696]: I0321 08:30:34.658866 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qvm82"] Mar 21 08:30:35 crc kubenswrapper[4696]: I0321 08:30:35.209788 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v9rlg" event={"ID":"85a5000b-3b68-4808-bcfc-93bcbde10dde","Type":"ContainerStarted","Data":"883148dde23dfd1d81101717644f6ddfa087aafb6d6f2a0659071c39c883dfbe"} Mar 21 08:30:35 crc kubenswrapper[4696]: I0321 08:30:35.212026 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jnw8m" event={"ID":"d70a46f0-1e5d-4edb-8745-9b8a41b66241","Type":"ContainerStarted","Data":"ad76337a74c84f6126a24fc0768cd55ff27a998a9d26ef3c78b7b12073b148c3"} Mar 21 08:30:35 crc kubenswrapper[4696]: I0321 08:30:35.213298 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"df078e35-cb4d-4be7-beaf-7fd6fc463a0c","Type":"ContainerStarted","Data":"fe1f72bf9ef983d3676ebee521a1f3b61709c61c083f362ba4402127ced1fb2f"} Mar 21 08:30:35 crc kubenswrapper[4696]: I0321 08:30:35.213344 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"df078e35-cb4d-4be7-beaf-7fd6fc463a0c","Type":"ContainerStarted","Data":"71115f4b3ed137978a13e46957c8c7f2ce7791599440e820b2edbda857c2b65b"} Mar 21 08:30:35 crc kubenswrapper[4696]: I0321 08:30:35.214623 4696 generic.go:334] "Generic (PLEG): container finished" podID="734c53f5-4c0a-4948-aa09-b8c20ce77920" containerID="e9f8daa6ce10c6a2ec2252195c75f71df55e35553e5827e0a484ca0708b80fb9" exitCode=0 Mar 21 08:30:35 crc kubenswrapper[4696]: I0321 08:30:35.214756 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"734c53f5-4c0a-4948-aa09-b8c20ce77920","Type":"ContainerDied","Data":"e9f8daa6ce10c6a2ec2252195c75f71df55e35553e5827e0a484ca0708b80fb9"} Mar 21 08:30:35 crc kubenswrapper[4696]: I0321 08:30:35.216798 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7wvkn" event={"ID":"0b88fc0a-dc29-4aa2-b625-c24162c0dc0b","Type":"ContainerStarted","Data":"6b4807b532396938086e91432b5e6318f6a16b97f2e22efd9dba5316a4b19196"} Mar 21 08:30:35 crc kubenswrapper[4696]: I0321 08:30:35.235020 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v9rlg" podStartSLOduration=4.778949341 podStartE2EDuration="46.235003898s" podCreationTimestamp="2026-03-21 08:29:49 +0000 UTC" firstStartedPulling="2026-03-21 08:29:53.205180892 +0000 UTC m=+127.326061605" lastFinishedPulling="2026-03-21 08:30:34.661235449 +0000 UTC m=+168.782116162" observedRunningTime="2026-03-21 08:30:35.23244975 +0000 UTC m=+169.353330483" watchObservedRunningTime="2026-03-21 08:30:35.235003898 +0000 UTC m=+169.355884611" Mar 21 08:30:35 crc kubenswrapper[4696]: I0321 08:30:35.251500 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jnw8m" podStartSLOduration=4.35839867 podStartE2EDuration="44.251486165s" podCreationTimestamp="2026-03-21 08:29:51 +0000 UTC" firstStartedPulling="2026-03-21 08:29:54.694692572 +0000 UTC m=+128.815573285" lastFinishedPulling="2026-03-21 08:30:34.587780067 +0000 UTC m=+168.708660780" observedRunningTime="2026-03-21 08:30:35.250314045 +0000 UTC m=+169.371194758" watchObservedRunningTime="2026-03-21 08:30:35.251486165 +0000 UTC m=+169.372366878" Mar 21 08:30:35 crc kubenswrapper[4696]: I0321 08:30:35.283519 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.283496806 podStartE2EDuration="2.283496806s" podCreationTimestamp="2026-03-21 08:30:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:30:35.282378307 +0000 UTC m=+169.403259020" watchObservedRunningTime="2026-03-21 08:30:35.283496806 +0000 UTC m=+169.404377519" Mar 21 08:30:35 crc kubenswrapper[4696]: I0321 08:30:35.308017 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7wvkn" podStartSLOduration=4.76888951 podStartE2EDuration="46.307999378s" podCreationTimestamp="2026-03-21 08:29:49 +0000 UTC" firstStartedPulling="2026-03-21 08:29:53.301461167 +0000 UTC m=+127.422341880" lastFinishedPulling="2026-03-21 08:30:34.840571035 +0000 UTC m=+168.961451748" observedRunningTime="2026-03-21 08:30:35.303651842 +0000 UTC m=+169.424532555" watchObservedRunningTime="2026-03-21 08:30:35.307999378 +0000 UTC m=+169.428880101" Mar 21 08:30:36 crc kubenswrapper[4696]: I0321 08:30:36.546943 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 21 08:30:36 crc kubenswrapper[4696]: I0321 08:30:36.740545 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/734c53f5-4c0a-4948-aa09-b8c20ce77920-kubelet-dir\") pod \"734c53f5-4c0a-4948-aa09-b8c20ce77920\" (UID: \"734c53f5-4c0a-4948-aa09-b8c20ce77920\") " Mar 21 08:30:36 crc kubenswrapper[4696]: I0321 08:30:36.740855 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/734c53f5-4c0a-4948-aa09-b8c20ce77920-kube-api-access\") pod \"734c53f5-4c0a-4948-aa09-b8c20ce77920\" (UID: \"734c53f5-4c0a-4948-aa09-b8c20ce77920\") " Mar 21 08:30:36 crc kubenswrapper[4696]: I0321 08:30:36.741880 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/734c53f5-4c0a-4948-aa09-b8c20ce77920-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "734c53f5-4c0a-4948-aa09-b8c20ce77920" (UID: "734c53f5-4c0a-4948-aa09-b8c20ce77920"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:30:36 crc kubenswrapper[4696]: I0321 08:30:36.746977 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/734c53f5-4c0a-4948-aa09-b8c20ce77920-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "734c53f5-4c0a-4948-aa09-b8c20ce77920" (UID: "734c53f5-4c0a-4948-aa09-b8c20ce77920"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:30:36 crc kubenswrapper[4696]: I0321 08:30:36.842664 4696 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/734c53f5-4c0a-4948-aa09-b8c20ce77920-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:36 crc kubenswrapper[4696]: I0321 08:30:36.842697 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/734c53f5-4c0a-4948-aa09-b8c20ce77920-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:37 crc kubenswrapper[4696]: I0321 08:30:37.227674 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"734c53f5-4c0a-4948-aa09-b8c20ce77920","Type":"ContainerDied","Data":"5157dddcd4907f20df02d18b5f7e04bac2f2901177dcfa6c800d6f0b732cec94"} Mar 21 08:30:37 crc kubenswrapper[4696]: I0321 08:30:37.227715 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5157dddcd4907f20df02d18b5f7e04bac2f2901177dcfa6c800d6f0b732cec94" Mar 21 08:30:37 crc kubenswrapper[4696]: I0321 08:30:37.227743 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 21 08:30:40 crc kubenswrapper[4696]: I0321 08:30:40.540703 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v9rlg" Mar 21 08:30:40 crc kubenswrapper[4696]: I0321 08:30:40.541491 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v9rlg" Mar 21 08:30:40 crc kubenswrapper[4696]: I0321 08:30:40.574721 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7wvkn" Mar 21 08:30:40 crc kubenswrapper[4696]: I0321 08:30:40.574770 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7wvkn" Mar 21 08:30:40 crc kubenswrapper[4696]: I0321 08:30:40.685228 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v9rlg" Mar 21 08:30:40 crc kubenswrapper[4696]: I0321 08:30:40.685643 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7wvkn" Mar 21 08:30:41 crc kubenswrapper[4696]: I0321 08:30:41.299535 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7wvkn" Mar 21 08:30:41 crc kubenswrapper[4696]: I0321 08:30:41.308780 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v9rlg" Mar 21 08:30:41 crc kubenswrapper[4696]: I0321 08:30:41.917353 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7wvkn"] Mar 21 08:30:42 crc kubenswrapper[4696]: I0321 08:30:42.071704 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jnw8m" Mar 21 08:30:42 crc kubenswrapper[4696]: I0321 08:30:42.071755 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jnw8m" Mar 21 08:30:42 crc kubenswrapper[4696]: I0321 08:30:42.107948 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jnw8m" Mar 21 08:30:42 crc kubenswrapper[4696]: I0321 08:30:42.290725 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jnw8m" Mar 21 08:30:43 crc kubenswrapper[4696]: I0321 08:30:43.255295 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7wvkn" podUID="0b88fc0a-dc29-4aa2-b625-c24162c0dc0b" containerName="registry-server" containerID="cri-o://6b4807b532396938086e91432b5e6318f6a16b97f2e22efd9dba5316a4b19196" gracePeriod=2 Mar 21 08:30:43 crc kubenswrapper[4696]: I0321 08:30:43.803972 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7wvkn" Mar 21 08:30:43 crc kubenswrapper[4696]: I0321 08:30:43.989673 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jth8k\" (UniqueName: \"kubernetes.io/projected/0b88fc0a-dc29-4aa2-b625-c24162c0dc0b-kube-api-access-jth8k\") pod \"0b88fc0a-dc29-4aa2-b625-c24162c0dc0b\" (UID: \"0b88fc0a-dc29-4aa2-b625-c24162c0dc0b\") " Mar 21 08:30:43 crc kubenswrapper[4696]: I0321 08:30:43.989774 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b88fc0a-dc29-4aa2-b625-c24162c0dc0b-catalog-content\") pod \"0b88fc0a-dc29-4aa2-b625-c24162c0dc0b\" (UID: \"0b88fc0a-dc29-4aa2-b625-c24162c0dc0b\") " Mar 21 08:30:43 crc kubenswrapper[4696]: I0321 08:30:43.989884 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b88fc0a-dc29-4aa2-b625-c24162c0dc0b-utilities\") pod \"0b88fc0a-dc29-4aa2-b625-c24162c0dc0b\" (UID: \"0b88fc0a-dc29-4aa2-b625-c24162c0dc0b\") " Mar 21 08:30:43 crc kubenswrapper[4696]: I0321 08:30:43.991000 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b88fc0a-dc29-4aa2-b625-c24162c0dc0b-utilities" (OuterVolumeSpecName: "utilities") pod "0b88fc0a-dc29-4aa2-b625-c24162c0dc0b" (UID: "0b88fc0a-dc29-4aa2-b625-c24162c0dc0b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:30:43 crc kubenswrapper[4696]: I0321 08:30:43.998211 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b88fc0a-dc29-4aa2-b625-c24162c0dc0b-kube-api-access-jth8k" (OuterVolumeSpecName: "kube-api-access-jth8k") pod "0b88fc0a-dc29-4aa2-b625-c24162c0dc0b" (UID: "0b88fc0a-dc29-4aa2-b625-c24162c0dc0b"). InnerVolumeSpecName "kube-api-access-jth8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.066138 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b88fc0a-dc29-4aa2-b625-c24162c0dc0b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0b88fc0a-dc29-4aa2-b625-c24162c0dc0b" (UID: "0b88fc0a-dc29-4aa2-b625-c24162c0dc0b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.091973 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jth8k\" (UniqueName: \"kubernetes.io/projected/0b88fc0a-dc29-4aa2-b625-c24162c0dc0b-kube-api-access-jth8k\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.092018 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b88fc0a-dc29-4aa2-b625-c24162c0dc0b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.092029 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b88fc0a-dc29-4aa2-b625-c24162c0dc0b-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.261743 4696 generic.go:334] "Generic (PLEG): container finished" podID="0030cc8a-f473-4d0a-ace9-728b66e62634" containerID="a07537b48c01840383ea0d043cf82e3aaf698087df27576814f170898198a1ad" exitCode=0 Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.261833 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cfjs8" event={"ID":"0030cc8a-f473-4d0a-ace9-728b66e62634","Type":"ContainerDied","Data":"a07537b48c01840383ea0d043cf82e3aaf698087df27576814f170898198a1ad"} Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.266802 4696 generic.go:334] "Generic (PLEG): container finished" podID="d6b6d345-41c3-4926-bf33-bac67ac54755" containerID="4566194647ac029d7b2db7bbd09a1d36de0d0bec6e2bb6fe2d5ee7b0eba7fc9d" exitCode=0 Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.266895 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trvd2" event={"ID":"d6b6d345-41c3-4926-bf33-bac67ac54755","Type":"ContainerDied","Data":"4566194647ac029d7b2db7bbd09a1d36de0d0bec6e2bb6fe2d5ee7b0eba7fc9d"} Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.268704 4696 generic.go:334] "Generic (PLEG): container finished" podID="0b88fc0a-dc29-4aa2-b625-c24162c0dc0b" containerID="6b4807b532396938086e91432b5e6318f6a16b97f2e22efd9dba5316a4b19196" exitCode=0 Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.268746 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7wvkn" Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.268741 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7wvkn" event={"ID":"0b88fc0a-dc29-4aa2-b625-c24162c0dc0b","Type":"ContainerDied","Data":"6b4807b532396938086e91432b5e6318f6a16b97f2e22efd9dba5316a4b19196"} Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.268907 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7wvkn" event={"ID":"0b88fc0a-dc29-4aa2-b625-c24162c0dc0b","Type":"ContainerDied","Data":"89bc54fdd0a2fe3ec8e94d47ce3b3f8409f69e12d70f294b9e6c516115ebbb77"} Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.268929 4696 scope.go:117] "RemoveContainer" containerID="6b4807b532396938086e91432b5e6318f6a16b97f2e22efd9dba5316a4b19196" Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.284698 4696 scope.go:117] "RemoveContainer" containerID="f8e612f0cdd24c6280d8c802539db297637f3450bc73471405923922ea03ed0d" Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.303993 4696 scope.go:117] "RemoveContainer" containerID="bd58cdcb5e5ee859ecbd932d8fc0256b43c576c5301c7787d43ff6b4ef9a684b" Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.318410 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7wvkn"] Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.322628 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7wvkn"] Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.330775 4696 scope.go:117] "RemoveContainer" containerID="6b4807b532396938086e91432b5e6318f6a16b97f2e22efd9dba5316a4b19196" Mar 21 08:30:44 crc kubenswrapper[4696]: E0321 08:30:44.331329 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b4807b532396938086e91432b5e6318f6a16b97f2e22efd9dba5316a4b19196\": container with ID starting with 6b4807b532396938086e91432b5e6318f6a16b97f2e22efd9dba5316a4b19196 not found: ID does not exist" containerID="6b4807b532396938086e91432b5e6318f6a16b97f2e22efd9dba5316a4b19196" Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.331362 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b4807b532396938086e91432b5e6318f6a16b97f2e22efd9dba5316a4b19196"} err="failed to get container status \"6b4807b532396938086e91432b5e6318f6a16b97f2e22efd9dba5316a4b19196\": rpc error: code = NotFound desc = could not find container \"6b4807b532396938086e91432b5e6318f6a16b97f2e22efd9dba5316a4b19196\": container with ID starting with 6b4807b532396938086e91432b5e6318f6a16b97f2e22efd9dba5316a4b19196 not found: ID does not exist" Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.331393 4696 scope.go:117] "RemoveContainer" containerID="f8e612f0cdd24c6280d8c802539db297637f3450bc73471405923922ea03ed0d" Mar 21 08:30:44 crc kubenswrapper[4696]: E0321 08:30:44.331707 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8e612f0cdd24c6280d8c802539db297637f3450bc73471405923922ea03ed0d\": container with ID starting with f8e612f0cdd24c6280d8c802539db297637f3450bc73471405923922ea03ed0d not found: ID does not exist" containerID="f8e612f0cdd24c6280d8c802539db297637f3450bc73471405923922ea03ed0d" Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.331733 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8e612f0cdd24c6280d8c802539db297637f3450bc73471405923922ea03ed0d"} err="failed to get container status \"f8e612f0cdd24c6280d8c802539db297637f3450bc73471405923922ea03ed0d\": rpc error: code = NotFound desc = could not find container \"f8e612f0cdd24c6280d8c802539db297637f3450bc73471405923922ea03ed0d\": container with ID starting with f8e612f0cdd24c6280d8c802539db297637f3450bc73471405923922ea03ed0d not found: ID does not exist" Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.331751 4696 scope.go:117] "RemoveContainer" containerID="bd58cdcb5e5ee859ecbd932d8fc0256b43c576c5301c7787d43ff6b4ef9a684b" Mar 21 08:30:44 crc kubenswrapper[4696]: E0321 08:30:44.332142 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd58cdcb5e5ee859ecbd932d8fc0256b43c576c5301c7787d43ff6b4ef9a684b\": container with ID starting with bd58cdcb5e5ee859ecbd932d8fc0256b43c576c5301c7787d43ff6b4ef9a684b not found: ID does not exist" containerID="bd58cdcb5e5ee859ecbd932d8fc0256b43c576c5301c7787d43ff6b4ef9a684b" Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.332161 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd58cdcb5e5ee859ecbd932d8fc0256b43c576c5301c7787d43ff6b4ef9a684b"} err="failed to get container status \"bd58cdcb5e5ee859ecbd932d8fc0256b43c576c5301c7787d43ff6b4ef9a684b\": rpc error: code = NotFound desc = could not find container \"bd58cdcb5e5ee859ecbd932d8fc0256b43c576c5301c7787d43ff6b4ef9a684b\": container with ID starting with bd58cdcb5e5ee859ecbd932d8fc0256b43c576c5301c7787d43ff6b4ef9a684b not found: ID does not exist" Mar 21 08:30:44 crc kubenswrapper[4696]: I0321 08:30:44.552044 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b88fc0a-dc29-4aa2-b625-c24162c0dc0b" path="/var/lib/kubelet/pods/0b88fc0a-dc29-4aa2-b625-c24162c0dc0b/volumes" Mar 21 08:30:45 crc kubenswrapper[4696]: I0321 08:30:45.484417 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6"] Mar 21 08:30:45 crc kubenswrapper[4696]: I0321 08:30:45.484902 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" podUID="e77a6440-cb75-4a20-bc27-ee519003841b" containerName="controller-manager" containerID="cri-o://1d832e1b53602f59f869850faadcbfdb0912be4ba51e640d93b76c75221d9122" gracePeriod=30 Mar 21 08:30:45 crc kubenswrapper[4696]: I0321 08:30:45.500168 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt"] Mar 21 08:30:45 crc kubenswrapper[4696]: I0321 08:30:45.500469 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" podUID="d6fa2cee-b7db-4516-997f-1602ecf67afb" containerName="route-controller-manager" containerID="cri-o://1b3fe6812a90966c6ced0db1da6ee7f83691cb0d2661624483dc1ce8bae0eef1" gracePeriod=30 Mar 21 08:30:45 crc kubenswrapper[4696]: I0321 08:30:45.990245 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.000130 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.118505 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrwvf\" (UniqueName: \"kubernetes.io/projected/e77a6440-cb75-4a20-bc27-ee519003841b-kube-api-access-wrwvf\") pod \"e77a6440-cb75-4a20-bc27-ee519003841b\" (UID: \"e77a6440-cb75-4a20-bc27-ee519003841b\") " Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.118548 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d6fa2cee-b7db-4516-997f-1602ecf67afb-client-ca\") pod \"d6fa2cee-b7db-4516-997f-1602ecf67afb\" (UID: \"d6fa2cee-b7db-4516-997f-1602ecf67afb\") " Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.118585 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e77a6440-cb75-4a20-bc27-ee519003841b-config\") pod \"e77a6440-cb75-4a20-bc27-ee519003841b\" (UID: \"e77a6440-cb75-4a20-bc27-ee519003841b\") " Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.118606 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e77a6440-cb75-4a20-bc27-ee519003841b-client-ca\") pod \"e77a6440-cb75-4a20-bc27-ee519003841b\" (UID: \"e77a6440-cb75-4a20-bc27-ee519003841b\") " Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.118622 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e77a6440-cb75-4a20-bc27-ee519003841b-proxy-ca-bundles\") pod \"e77a6440-cb75-4a20-bc27-ee519003841b\" (UID: \"e77a6440-cb75-4a20-bc27-ee519003841b\") " Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.118662 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e77a6440-cb75-4a20-bc27-ee519003841b-serving-cert\") pod \"e77a6440-cb75-4a20-bc27-ee519003841b\" (UID: \"e77a6440-cb75-4a20-bc27-ee519003841b\") " Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.118684 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6fa2cee-b7db-4516-997f-1602ecf67afb-config\") pod \"d6fa2cee-b7db-4516-997f-1602ecf67afb\" (UID: \"d6fa2cee-b7db-4516-997f-1602ecf67afb\") " Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.118734 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6fa2cee-b7db-4516-997f-1602ecf67afb-serving-cert\") pod \"d6fa2cee-b7db-4516-997f-1602ecf67afb\" (UID: \"d6fa2cee-b7db-4516-997f-1602ecf67afb\") " Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.118791 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ntrb\" (UniqueName: \"kubernetes.io/projected/d6fa2cee-b7db-4516-997f-1602ecf67afb-kube-api-access-5ntrb\") pod \"d6fa2cee-b7db-4516-997f-1602ecf67afb\" (UID: \"d6fa2cee-b7db-4516-997f-1602ecf67afb\") " Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.119910 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6fa2cee-b7db-4516-997f-1602ecf67afb-client-ca" (OuterVolumeSpecName: "client-ca") pod "d6fa2cee-b7db-4516-997f-1602ecf67afb" (UID: "d6fa2cee-b7db-4516-997f-1602ecf67afb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.119991 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6fa2cee-b7db-4516-997f-1602ecf67afb-config" (OuterVolumeSpecName: "config") pod "d6fa2cee-b7db-4516-997f-1602ecf67afb" (UID: "d6fa2cee-b7db-4516-997f-1602ecf67afb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.120161 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e77a6440-cb75-4a20-bc27-ee519003841b-config" (OuterVolumeSpecName: "config") pod "e77a6440-cb75-4a20-bc27-ee519003841b" (UID: "e77a6440-cb75-4a20-bc27-ee519003841b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.120248 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e77a6440-cb75-4a20-bc27-ee519003841b-client-ca" (OuterVolumeSpecName: "client-ca") pod "e77a6440-cb75-4a20-bc27-ee519003841b" (UID: "e77a6440-cb75-4a20-bc27-ee519003841b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.120375 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e77a6440-cb75-4a20-bc27-ee519003841b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e77a6440-cb75-4a20-bc27-ee519003841b" (UID: "e77a6440-cb75-4a20-bc27-ee519003841b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.124050 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e77a6440-cb75-4a20-bc27-ee519003841b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e77a6440-cb75-4a20-bc27-ee519003841b" (UID: "e77a6440-cb75-4a20-bc27-ee519003841b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.124214 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6fa2cee-b7db-4516-997f-1602ecf67afb-kube-api-access-5ntrb" (OuterVolumeSpecName: "kube-api-access-5ntrb") pod "d6fa2cee-b7db-4516-997f-1602ecf67afb" (UID: "d6fa2cee-b7db-4516-997f-1602ecf67afb"). InnerVolumeSpecName "kube-api-access-5ntrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.124517 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6fa2cee-b7db-4516-997f-1602ecf67afb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d6fa2cee-b7db-4516-997f-1602ecf67afb" (UID: "d6fa2cee-b7db-4516-997f-1602ecf67afb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.124925 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e77a6440-cb75-4a20-bc27-ee519003841b-kube-api-access-wrwvf" (OuterVolumeSpecName: "kube-api-access-wrwvf") pod "e77a6440-cb75-4a20-bc27-ee519003841b" (UID: "e77a6440-cb75-4a20-bc27-ee519003841b"). InnerVolumeSpecName "kube-api-access-wrwvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.219944 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6fa2cee-b7db-4516-997f-1602ecf67afb-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.219985 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ntrb\" (UniqueName: \"kubernetes.io/projected/d6fa2cee-b7db-4516-997f-1602ecf67afb-kube-api-access-5ntrb\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.219998 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrwvf\" (UniqueName: \"kubernetes.io/projected/e77a6440-cb75-4a20-bc27-ee519003841b-kube-api-access-wrwvf\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.220009 4696 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d6fa2cee-b7db-4516-997f-1602ecf67afb-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.220021 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e77a6440-cb75-4a20-bc27-ee519003841b-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.220028 4696 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e77a6440-cb75-4a20-bc27-ee519003841b-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.220037 4696 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e77a6440-cb75-4a20-bc27-ee519003841b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.220045 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e77a6440-cb75-4a20-bc27-ee519003841b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.220054 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6fa2cee-b7db-4516-997f-1602ecf67afb-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.281228 4696 generic.go:334] "Generic (PLEG): container finished" podID="d6fa2cee-b7db-4516-997f-1602ecf67afb" containerID="1b3fe6812a90966c6ced0db1da6ee7f83691cb0d2661624483dc1ce8bae0eef1" exitCode=0 Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.281308 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" event={"ID":"d6fa2cee-b7db-4516-997f-1602ecf67afb","Type":"ContainerDied","Data":"1b3fe6812a90966c6ced0db1da6ee7f83691cb0d2661624483dc1ce8bae0eef1"} Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.281319 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.281339 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt" event={"ID":"d6fa2cee-b7db-4516-997f-1602ecf67afb","Type":"ContainerDied","Data":"2147be457b9cb2b43807422306b17df406016caa106ea12320ef643120523568"} Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.281357 4696 scope.go:117] "RemoveContainer" containerID="1b3fe6812a90966c6ced0db1da6ee7f83691cb0d2661624483dc1ce8bae0eef1" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.285758 4696 generic.go:334] "Generic (PLEG): container finished" podID="5149cc5e-d877-4d52-8db6-7df6ae060ab0" containerID="dd22af6e40a86b54aa68312be29327f0ab663e8f67b7db7c3187a1d823a0c9ae" exitCode=0 Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.286221 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7tpgm" event={"ID":"5149cc5e-d877-4d52-8db6-7df6ae060ab0","Type":"ContainerDied","Data":"dd22af6e40a86b54aa68312be29327f0ab663e8f67b7db7c3187a1d823a0c9ae"} Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.289500 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cfjs8" event={"ID":"0030cc8a-f473-4d0a-ace9-728b66e62634","Type":"ContainerStarted","Data":"3ce447a4777038b8616bc6dd26064b084fa129af52ead66b5221e39c940bdc0c"} Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.293402 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trvd2" event={"ID":"d6b6d345-41c3-4926-bf33-bac67ac54755","Type":"ContainerStarted","Data":"8effc5cfb7634b8f392e55cd369d2a4d9310dce98fa9bd6a01e981736f200fb5"} Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.298805 4696 generic.go:334] "Generic (PLEG): container finished" podID="e77a6440-cb75-4a20-bc27-ee519003841b" containerID="1d832e1b53602f59f869850faadcbfdb0912be4ba51e640d93b76c75221d9122" exitCode=0 Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.298931 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" event={"ID":"e77a6440-cb75-4a20-bc27-ee519003841b","Type":"ContainerDied","Data":"1d832e1b53602f59f869850faadcbfdb0912be4ba51e640d93b76c75221d9122"} Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.299028 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" event={"ID":"e77a6440-cb75-4a20-bc27-ee519003841b","Type":"ContainerDied","Data":"2deafa2e79cddd590771b15bb45d6e740040bea32ae603d9ee3649c6dff2eb29"} Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.299052 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.302417 4696 scope.go:117] "RemoveContainer" containerID="1b3fe6812a90966c6ced0db1da6ee7f83691cb0d2661624483dc1ce8bae0eef1" Mar 21 08:30:46 crc kubenswrapper[4696]: E0321 08:30:46.303025 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b3fe6812a90966c6ced0db1da6ee7f83691cb0d2661624483dc1ce8bae0eef1\": container with ID starting with 1b3fe6812a90966c6ced0db1da6ee7f83691cb0d2661624483dc1ce8bae0eef1 not found: ID does not exist" containerID="1b3fe6812a90966c6ced0db1da6ee7f83691cb0d2661624483dc1ce8bae0eef1" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.303065 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b3fe6812a90966c6ced0db1da6ee7f83691cb0d2661624483dc1ce8bae0eef1"} err="failed to get container status \"1b3fe6812a90966c6ced0db1da6ee7f83691cb0d2661624483dc1ce8bae0eef1\": rpc error: code = NotFound desc = could not find container \"1b3fe6812a90966c6ced0db1da6ee7f83691cb0d2661624483dc1ce8bae0eef1\": container with ID starting with 1b3fe6812a90966c6ced0db1da6ee7f83691cb0d2661624483dc1ce8bae0eef1 not found: ID does not exist" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.303114 4696 scope.go:117] "RemoveContainer" containerID="1d832e1b53602f59f869850faadcbfdb0912be4ba51e640d93b76c75221d9122" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.323706 4696 scope.go:117] "RemoveContainer" containerID="1d832e1b53602f59f869850faadcbfdb0912be4ba51e640d93b76c75221d9122" Mar 21 08:30:46 crc kubenswrapper[4696]: E0321 08:30:46.325306 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d832e1b53602f59f869850faadcbfdb0912be4ba51e640d93b76c75221d9122\": container with ID starting with 1d832e1b53602f59f869850faadcbfdb0912be4ba51e640d93b76c75221d9122 not found: ID does not exist" containerID="1d832e1b53602f59f869850faadcbfdb0912be4ba51e640d93b76c75221d9122" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.325342 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d832e1b53602f59f869850faadcbfdb0912be4ba51e640d93b76c75221d9122"} err="failed to get container status \"1d832e1b53602f59f869850faadcbfdb0912be4ba51e640d93b76c75221d9122\": rpc error: code = NotFound desc = could not find container \"1d832e1b53602f59f869850faadcbfdb0912be4ba51e640d93b76c75221d9122\": container with ID starting with 1d832e1b53602f59f869850faadcbfdb0912be4ba51e640d93b76c75221d9122 not found: ID does not exist" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.334392 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cfjs8" podStartSLOduration=2.9846513850000003 podStartE2EDuration="55.334370897s" podCreationTimestamp="2026-03-21 08:29:51 +0000 UTC" firstStartedPulling="2026-03-21 08:29:53.494878109 +0000 UTC m=+127.615758822" lastFinishedPulling="2026-03-21 08:30:45.844597621 +0000 UTC m=+179.965478334" observedRunningTime="2026-03-21 08:30:46.330670249 +0000 UTC m=+180.451550992" watchObservedRunningTime="2026-03-21 08:30:46.334370897 +0000 UTC m=+180.455251610" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.346176 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt"] Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.349423 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74dfdbbfc5-kxrnt"] Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.368248 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-trvd2" podStartSLOduration=3.485697989 podStartE2EDuration="54.368229157s" podCreationTimestamp="2026-03-21 08:29:52 +0000 UTC" firstStartedPulling="2026-03-21 08:29:54.632984867 +0000 UTC m=+128.753865580" lastFinishedPulling="2026-03-21 08:30:45.515516035 +0000 UTC m=+179.636396748" observedRunningTime="2026-03-21 08:30:46.366698396 +0000 UTC m=+180.487579109" watchObservedRunningTime="2026-03-21 08:30:46.368229157 +0000 UTC m=+180.489109860" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.388181 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6"] Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.389472 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7dbddf9fd5-td2n6"] Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.547744 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6fa2cee-b7db-4516-997f-1602ecf67afb" path="/var/lib/kubelet/pods/d6fa2cee-b7db-4516-997f-1602ecf67afb/volumes" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.548346 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e77a6440-cb75-4a20-bc27-ee519003841b" path="/var/lib/kubelet/pods/e77a6440-cb75-4a20-bc27-ee519003841b/volumes" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.928630 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7dd5d7c474-spgtg"] Mar 21 08:30:46 crc kubenswrapper[4696]: E0321 08:30:46.929264 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6fa2cee-b7db-4516-997f-1602ecf67afb" containerName="route-controller-manager" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.929286 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6fa2cee-b7db-4516-997f-1602ecf67afb" containerName="route-controller-manager" Mar 21 08:30:46 crc kubenswrapper[4696]: E0321 08:30:46.929300 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b88fc0a-dc29-4aa2-b625-c24162c0dc0b" containerName="registry-server" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.929308 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b88fc0a-dc29-4aa2-b625-c24162c0dc0b" containerName="registry-server" Mar 21 08:30:46 crc kubenswrapper[4696]: E0321 08:30:46.929319 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b88fc0a-dc29-4aa2-b625-c24162c0dc0b" containerName="extract-utilities" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.929326 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b88fc0a-dc29-4aa2-b625-c24162c0dc0b" containerName="extract-utilities" Mar 21 08:30:46 crc kubenswrapper[4696]: E0321 08:30:46.929339 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e77a6440-cb75-4a20-bc27-ee519003841b" containerName="controller-manager" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.929346 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e77a6440-cb75-4a20-bc27-ee519003841b" containerName="controller-manager" Mar 21 08:30:46 crc kubenswrapper[4696]: E0321 08:30:46.929364 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b88fc0a-dc29-4aa2-b625-c24162c0dc0b" containerName="extract-content" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.929370 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b88fc0a-dc29-4aa2-b625-c24162c0dc0b" containerName="extract-content" Mar 21 08:30:46 crc kubenswrapper[4696]: E0321 08:30:46.929379 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="734c53f5-4c0a-4948-aa09-b8c20ce77920" containerName="pruner" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.929386 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="734c53f5-4c0a-4948-aa09-b8c20ce77920" containerName="pruner" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.929519 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6fa2cee-b7db-4516-997f-1602ecf67afb" containerName="route-controller-manager" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.929538 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="734c53f5-4c0a-4948-aa09-b8c20ce77920" containerName="pruner" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.929546 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b88fc0a-dc29-4aa2-b625-c24162c0dc0b" containerName="registry-server" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.929553 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="e77a6440-cb75-4a20-bc27-ee519003841b" containerName="controller-manager" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.930044 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.932614 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj"] Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.933377 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.938018 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.942500 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.942762 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.942800 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.943312 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.945270 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.945575 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.945715 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.945761 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.946198 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.950530 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.955449 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.965046 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj"] Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.965700 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7dd5d7c474-spgtg"] Mar 21 08:30:46 crc kubenswrapper[4696]: I0321 08:30:46.967425 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.132286 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-config\") pod \"controller-manager-7dd5d7c474-spgtg\" (UID: \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\") " pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.132333 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp4nj\" (UniqueName: \"kubernetes.io/projected/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-kube-api-access-wp4nj\") pod \"route-controller-manager-5b75455d89-xwqgj\" (UID: \"fd34f488-8a79-4e8f-8fdf-baaeefc325ad\") " pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.132382 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-client-ca\") pod \"route-controller-manager-5b75455d89-xwqgj\" (UID: \"fd34f488-8a79-4e8f-8fdf-baaeefc325ad\") " pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.132412 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-proxy-ca-bundles\") pod \"controller-manager-7dd5d7c474-spgtg\" (UID: \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\") " pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.132436 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf2pd\" (UniqueName: \"kubernetes.io/projected/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-kube-api-access-zf2pd\") pod \"controller-manager-7dd5d7c474-spgtg\" (UID: \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\") " pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.132461 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-serving-cert\") pod \"route-controller-manager-5b75455d89-xwqgj\" (UID: \"fd34f488-8a79-4e8f-8fdf-baaeefc325ad\") " pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.132494 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-serving-cert\") pod \"controller-manager-7dd5d7c474-spgtg\" (UID: \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\") " pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.132529 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-config\") pod \"route-controller-manager-5b75455d89-xwqgj\" (UID: \"fd34f488-8a79-4e8f-8fdf-baaeefc325ad\") " pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.132569 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-client-ca\") pod \"controller-manager-7dd5d7c474-spgtg\" (UID: \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\") " pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.233533 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-serving-cert\") pod \"controller-manager-7dd5d7c474-spgtg\" (UID: \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\") " pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.233752 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-config\") pod \"route-controller-manager-5b75455d89-xwqgj\" (UID: \"fd34f488-8a79-4e8f-8fdf-baaeefc325ad\") " pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.233939 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-client-ca\") pod \"controller-manager-7dd5d7c474-spgtg\" (UID: \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\") " pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.234039 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-config\") pod \"controller-manager-7dd5d7c474-spgtg\" (UID: \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\") " pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.234127 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp4nj\" (UniqueName: \"kubernetes.io/projected/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-kube-api-access-wp4nj\") pod \"route-controller-manager-5b75455d89-xwqgj\" (UID: \"fd34f488-8a79-4e8f-8fdf-baaeefc325ad\") " pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.234250 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-client-ca\") pod \"route-controller-manager-5b75455d89-xwqgj\" (UID: \"fd34f488-8a79-4e8f-8fdf-baaeefc325ad\") " pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.234348 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-proxy-ca-bundles\") pod \"controller-manager-7dd5d7c474-spgtg\" (UID: \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\") " pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.234422 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zf2pd\" (UniqueName: \"kubernetes.io/projected/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-kube-api-access-zf2pd\") pod \"controller-manager-7dd5d7c474-spgtg\" (UID: \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\") " pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.234499 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-serving-cert\") pod \"route-controller-manager-5b75455d89-xwqgj\" (UID: \"fd34f488-8a79-4e8f-8fdf-baaeefc325ad\") " pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.235108 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-client-ca\") pod \"controller-manager-7dd5d7c474-spgtg\" (UID: \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\") " pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.235534 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-config\") pod \"controller-manager-7dd5d7c474-spgtg\" (UID: \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\") " pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.235574 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-client-ca\") pod \"route-controller-manager-5b75455d89-xwqgj\" (UID: \"fd34f488-8a79-4e8f-8fdf-baaeefc325ad\") " pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.235796 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-config\") pod \"route-controller-manager-5b75455d89-xwqgj\" (UID: \"fd34f488-8a79-4e8f-8fdf-baaeefc325ad\") " pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.236973 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-proxy-ca-bundles\") pod \"controller-manager-7dd5d7c474-spgtg\" (UID: \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\") " pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.238432 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-serving-cert\") pod \"route-controller-manager-5b75455d89-xwqgj\" (UID: \"fd34f488-8a79-4e8f-8fdf-baaeefc325ad\") " pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.238456 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-serving-cert\") pod \"controller-manager-7dd5d7c474-spgtg\" (UID: \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\") " pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.257546 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf2pd\" (UniqueName: \"kubernetes.io/projected/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-kube-api-access-zf2pd\") pod \"controller-manager-7dd5d7c474-spgtg\" (UID: \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\") " pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.257922 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.260968 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp4nj\" (UniqueName: \"kubernetes.io/projected/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-kube-api-access-wp4nj\") pod \"route-controller-manager-5b75455d89-xwqgj\" (UID: \"fd34f488-8a79-4e8f-8fdf-baaeefc325ad\") " pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.270120 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.313771 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7tpgm" event={"ID":"5149cc5e-d877-4d52-8db6-7df6ae060ab0","Type":"ContainerStarted","Data":"b51ab18a4addc9706badfaf2bcc61a96534b87c4bcabdd609307d475c963a487"} Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.316068 4696 generic.go:334] "Generic (PLEG): container finished" podID="6ed31bad-6fb5-49be-922b-27198867ca4a" containerID="2984cd8c015b136bc5bcc3860dd8c7f19b0a93e677a874539ef2aed10847c968" exitCode=0 Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.316090 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jmqr7" event={"ID":"6ed31bad-6fb5-49be-922b-27198867ca4a","Type":"ContainerDied","Data":"2984cd8c015b136bc5bcc3860dd8c7f19b0a93e677a874539ef2aed10847c968"} Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.332600 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7tpgm" podStartSLOduration=5.189823145 podStartE2EDuration="58.332582496s" podCreationTimestamp="2026-03-21 08:29:49 +0000 UTC" firstStartedPulling="2026-03-21 08:29:53.571072751 +0000 UTC m=+127.691953464" lastFinishedPulling="2026-03-21 08:30:46.713832102 +0000 UTC m=+180.834712815" observedRunningTime="2026-03-21 08:30:47.332310009 +0000 UTC m=+181.453190722" watchObservedRunningTime="2026-03-21 08:30:47.332582496 +0000 UTC m=+181.453463209" Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.553966 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj"] Mar 21 08:30:47 crc kubenswrapper[4696]: W0321 08:30:47.563686 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd34f488_8a79_4e8f_8fdf_baaeefc325ad.slice/crio-a379de360a98eb26a3ae56623f9803e071cf77d734691aa3e980a6715b4339c9 WatchSource:0}: Error finding container a379de360a98eb26a3ae56623f9803e071cf77d734691aa3e980a6715b4339c9: Status 404 returned error can't find the container with id a379de360a98eb26a3ae56623f9803e071cf77d734691aa3e980a6715b4339c9 Mar 21 08:30:47 crc kubenswrapper[4696]: I0321 08:30:47.696202 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7dd5d7c474-spgtg"] Mar 21 08:30:47 crc kubenswrapper[4696]: W0321 08:30:47.702513 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ce87a6d_6ef8_4c95_8f74_d1597371ae1a.slice/crio-86da31bb47d471661ddae8a74649e35d105ad0aed4ba2b92f7bc892cafdd591c WatchSource:0}: Error finding container 86da31bb47d471661ddae8a74649e35d105ad0aed4ba2b92f7bc892cafdd591c: Status 404 returned error can't find the container with id 86da31bb47d471661ddae8a74649e35d105ad0aed4ba2b92f7bc892cafdd591c Mar 21 08:30:48 crc kubenswrapper[4696]: I0321 08:30:48.322114 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" event={"ID":"fd34f488-8a79-4e8f-8fdf-baaeefc325ad","Type":"ContainerStarted","Data":"8939678ed5699cc486f5bd92aa356cf2e8ce9c133d16a3aa9893a3ba4a8f1c3c"} Mar 21 08:30:48 crc kubenswrapper[4696]: I0321 08:30:48.322760 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" Mar 21 08:30:48 crc kubenswrapper[4696]: I0321 08:30:48.322845 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" event={"ID":"fd34f488-8a79-4e8f-8fdf-baaeefc325ad","Type":"ContainerStarted","Data":"a379de360a98eb26a3ae56623f9803e071cf77d734691aa3e980a6715b4339c9"} Mar 21 08:30:48 crc kubenswrapper[4696]: I0321 08:30:48.323429 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" event={"ID":"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a","Type":"ContainerStarted","Data":"0e2c8238b924613669a5ddf1009989ad85ce102a5eb501a92e47f8bac56b0670"} Mar 21 08:30:48 crc kubenswrapper[4696]: I0321 08:30:48.323487 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" event={"ID":"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a","Type":"ContainerStarted","Data":"86da31bb47d471661ddae8a74649e35d105ad0aed4ba2b92f7bc892cafdd591c"} Mar 21 08:30:48 crc kubenswrapper[4696]: I0321 08:30:48.349012 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" podStartSLOduration=3.3489988889999998 podStartE2EDuration="3.348998889s" podCreationTimestamp="2026-03-21 08:30:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:30:48.347045386 +0000 UTC m=+182.467926099" watchObservedRunningTime="2026-03-21 08:30:48.348998889 +0000 UTC m=+182.469879602" Mar 21 08:30:48 crc kubenswrapper[4696]: I0321 08:30:48.545867 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" podStartSLOduration=3.54584836 podStartE2EDuration="3.54584836s" podCreationTimestamp="2026-03-21 08:30:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:30:48.380079535 +0000 UTC m=+182.500960248" watchObservedRunningTime="2026-03-21 08:30:48.54584836 +0000 UTC m=+182.666729073" Mar 21 08:30:48 crc kubenswrapper[4696]: I0321 08:30:48.673884 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" Mar 21 08:30:49 crc kubenswrapper[4696]: I0321 08:30:49.327806 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" Mar 21 08:30:49 crc kubenswrapper[4696]: I0321 08:30:49.332516 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" Mar 21 08:30:50 crc kubenswrapper[4696]: I0321 08:30:50.335571 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jmqr7" event={"ID":"6ed31bad-6fb5-49be-922b-27198867ca4a","Type":"ContainerStarted","Data":"2fb2abd4ababf21c2f49afb05b5050ad78ef3a7087ba8a4a1a5adc52ed8a1362"} Mar 21 08:30:50 crc kubenswrapper[4696]: I0321 08:30:50.356297 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jmqr7" podStartSLOduration=4.238083072 podStartE2EDuration="58.356276605s" podCreationTimestamp="2026-03-21 08:29:52 +0000 UTC" firstStartedPulling="2026-03-21 08:29:54.746062139 +0000 UTC m=+128.866942852" lastFinishedPulling="2026-03-21 08:30:48.864255672 +0000 UTC m=+182.985136385" observedRunningTime="2026-03-21 08:30:50.353356216 +0000 UTC m=+184.474236939" watchObservedRunningTime="2026-03-21 08:30:50.356276605 +0000 UTC m=+184.477157318" Mar 21 08:30:50 crc kubenswrapper[4696]: I0321 08:30:50.596959 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7tpgm" Mar 21 08:30:50 crc kubenswrapper[4696]: I0321 08:30:50.596999 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7tpgm" Mar 21 08:30:50 crc kubenswrapper[4696]: I0321 08:30:50.635856 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7tpgm" Mar 21 08:30:51 crc kubenswrapper[4696]: I0321 08:30:51.383568 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7tpgm" Mar 21 08:30:52 crc kubenswrapper[4696]: I0321 08:30:52.328788 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cfjs8" Mar 21 08:30:52 crc kubenswrapper[4696]: I0321 08:30:52.329972 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cfjs8" Mar 21 08:30:52 crc kubenswrapper[4696]: I0321 08:30:52.427754 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-trvd2" Mar 21 08:30:52 crc kubenswrapper[4696]: I0321 08:30:52.427836 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-trvd2" Mar 21 08:30:52 crc kubenswrapper[4696]: I0321 08:30:52.507576 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-trvd2" Mar 21 08:30:52 crc kubenswrapper[4696]: I0321 08:30:52.725755 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jmqr7" Mar 21 08:30:52 crc kubenswrapper[4696]: I0321 08:30:52.725802 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jmqr7" Mar 21 08:30:53 crc kubenswrapper[4696]: I0321 08:30:53.375283 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cfjs8" podUID="0030cc8a-f473-4d0a-ace9-728b66e62634" containerName="registry-server" probeResult="failure" output=< Mar 21 08:30:53 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Mar 21 08:30:53 crc kubenswrapper[4696]: > Mar 21 08:30:53 crc kubenswrapper[4696]: I0321 08:30:53.402514 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-trvd2" Mar 21 08:30:53 crc kubenswrapper[4696]: I0321 08:30:53.763344 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jmqr7" podUID="6ed31bad-6fb5-49be-922b-27198867ca4a" containerName="registry-server" probeResult="failure" output=< Mar 21 08:30:53 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Mar 21 08:30:53 crc kubenswrapper[4696]: > Mar 21 08:30:54 crc kubenswrapper[4696]: I0321 08:30:54.323405 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7tpgm"] Mar 21 08:30:54 crc kubenswrapper[4696]: I0321 08:30:54.323803 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7tpgm" podUID="5149cc5e-d877-4d52-8db6-7df6ae060ab0" containerName="registry-server" containerID="cri-o://b51ab18a4addc9706badfaf2bcc61a96534b87c4bcabdd609307d475c963a487" gracePeriod=2 Mar 21 08:30:55 crc kubenswrapper[4696]: I0321 08:30:55.845082 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7tpgm" Mar 21 08:30:55 crc kubenswrapper[4696]: I0321 08:30:55.968983 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5149cc5e-d877-4d52-8db6-7df6ae060ab0-catalog-content\") pod \"5149cc5e-d877-4d52-8db6-7df6ae060ab0\" (UID: \"5149cc5e-d877-4d52-8db6-7df6ae060ab0\") " Mar 21 08:30:55 crc kubenswrapper[4696]: I0321 08:30:55.969070 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gs9s\" (UniqueName: \"kubernetes.io/projected/5149cc5e-d877-4d52-8db6-7df6ae060ab0-kube-api-access-7gs9s\") pod \"5149cc5e-d877-4d52-8db6-7df6ae060ab0\" (UID: \"5149cc5e-d877-4d52-8db6-7df6ae060ab0\") " Mar 21 08:30:55 crc kubenswrapper[4696]: I0321 08:30:55.969139 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5149cc5e-d877-4d52-8db6-7df6ae060ab0-utilities\") pod \"5149cc5e-d877-4d52-8db6-7df6ae060ab0\" (UID: \"5149cc5e-d877-4d52-8db6-7df6ae060ab0\") " Mar 21 08:30:55 crc kubenswrapper[4696]: I0321 08:30:55.970165 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5149cc5e-d877-4d52-8db6-7df6ae060ab0-utilities" (OuterVolumeSpecName: "utilities") pod "5149cc5e-d877-4d52-8db6-7df6ae060ab0" (UID: "5149cc5e-d877-4d52-8db6-7df6ae060ab0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:30:55 crc kubenswrapper[4696]: I0321 08:30:55.974327 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5149cc5e-d877-4d52-8db6-7df6ae060ab0-kube-api-access-7gs9s" (OuterVolumeSpecName: "kube-api-access-7gs9s") pod "5149cc5e-d877-4d52-8db6-7df6ae060ab0" (UID: "5149cc5e-d877-4d52-8db6-7df6ae060ab0"). InnerVolumeSpecName "kube-api-access-7gs9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.026487 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5149cc5e-d877-4d52-8db6-7df6ae060ab0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5149cc5e-d877-4d52-8db6-7df6ae060ab0" (UID: "5149cc5e-d877-4d52-8db6-7df6ae060ab0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.070431 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5149cc5e-d877-4d52-8db6-7df6ae060ab0-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.070474 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5149cc5e-d877-4d52-8db6-7df6ae060ab0-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.070490 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gs9s\" (UniqueName: \"kubernetes.io/projected/5149cc5e-d877-4d52-8db6-7df6ae060ab0-kube-api-access-7gs9s\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.250177 4696 csr.go:261] certificate signing request csr-dv7wm is approved, waiting to be issued Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.256250 4696 csr.go:257] certificate signing request csr-dv7wm is issued Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.372493 4696 generic.go:334] "Generic (PLEG): container finished" podID="6aabaca2-7def-4550-bae1-9337b7c1002b" containerID="c41c5c0054693270a82a25c88765741367992244e855001bd0c5b6a6fa6345a9" exitCode=0 Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.372568 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qsg67" event={"ID":"6aabaca2-7def-4550-bae1-9337b7c1002b","Type":"ContainerDied","Data":"c41c5c0054693270a82a25c88765741367992244e855001bd0c5b6a6fa6345a9"} Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.375349 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568030-4x55c" event={"ID":"a6581056-7677-491f-8f34-5626b43b6e77","Type":"ContainerStarted","Data":"89cca28a71bf3998ea6e1b5417a80f0fb41aadfedf5f4beb2237ac7e4ee7ae9e"} Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.380264 4696 generic.go:334] "Generic (PLEG): container finished" podID="5149cc5e-d877-4d52-8db6-7df6ae060ab0" containerID="b51ab18a4addc9706badfaf2bcc61a96534b87c4bcabdd609307d475c963a487" exitCode=0 Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.380307 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7tpgm" event={"ID":"5149cc5e-d877-4d52-8db6-7df6ae060ab0","Type":"ContainerDied","Data":"b51ab18a4addc9706badfaf2bcc61a96534b87c4bcabdd609307d475c963a487"} Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.380362 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7tpgm" Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.380385 4696 scope.go:117] "RemoveContainer" containerID="b51ab18a4addc9706badfaf2bcc61a96534b87c4bcabdd609307d475c963a487" Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.380370 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7tpgm" event={"ID":"5149cc5e-d877-4d52-8db6-7df6ae060ab0","Type":"ContainerDied","Data":"1af20a97d04f7edceaca510842173094d663f9bc603c207f36aa171da3369185"} Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.396382 4696 scope.go:117] "RemoveContainer" containerID="dd22af6e40a86b54aa68312be29327f0ab663e8f67b7db7c3187a1d823a0c9ae" Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.413430 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7tpgm"] Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.425135 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7tpgm"] Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.425888 4696 scope.go:117] "RemoveContainer" containerID="60d54f52b8f62e3f02ae6a78c86e02e67f31dea89ae92fc551f09f878a24b04a" Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.433084 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29568030-4x55c" podStartSLOduration=1.880404897 podStartE2EDuration="56.433071363s" podCreationTimestamp="2026-03-21 08:30:00 +0000 UTC" firstStartedPulling="2026-03-21 08:30:00.966811317 +0000 UTC m=+135.087692030" lastFinishedPulling="2026-03-21 08:30:55.519477753 +0000 UTC m=+189.640358496" observedRunningTime="2026-03-21 08:30:56.430144236 +0000 UTC m=+190.551024949" watchObservedRunningTime="2026-03-21 08:30:56.433071363 +0000 UTC m=+190.553952076" Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.441852 4696 scope.go:117] "RemoveContainer" containerID="b51ab18a4addc9706badfaf2bcc61a96534b87c4bcabdd609307d475c963a487" Mar 21 08:30:56 crc kubenswrapper[4696]: E0321 08:30:56.443252 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b51ab18a4addc9706badfaf2bcc61a96534b87c4bcabdd609307d475c963a487\": container with ID starting with b51ab18a4addc9706badfaf2bcc61a96534b87c4bcabdd609307d475c963a487 not found: ID does not exist" containerID="b51ab18a4addc9706badfaf2bcc61a96534b87c4bcabdd609307d475c963a487" Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.443359 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b51ab18a4addc9706badfaf2bcc61a96534b87c4bcabdd609307d475c963a487"} err="failed to get container status \"b51ab18a4addc9706badfaf2bcc61a96534b87c4bcabdd609307d475c963a487\": rpc error: code = NotFound desc = could not find container \"b51ab18a4addc9706badfaf2bcc61a96534b87c4bcabdd609307d475c963a487\": container with ID starting with b51ab18a4addc9706badfaf2bcc61a96534b87c4bcabdd609307d475c963a487 not found: ID does not exist" Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.443450 4696 scope.go:117] "RemoveContainer" containerID="dd22af6e40a86b54aa68312be29327f0ab663e8f67b7db7c3187a1d823a0c9ae" Mar 21 08:30:56 crc kubenswrapper[4696]: E0321 08:30:56.443930 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd22af6e40a86b54aa68312be29327f0ab663e8f67b7db7c3187a1d823a0c9ae\": container with ID starting with dd22af6e40a86b54aa68312be29327f0ab663e8f67b7db7c3187a1d823a0c9ae not found: ID does not exist" containerID="dd22af6e40a86b54aa68312be29327f0ab663e8f67b7db7c3187a1d823a0c9ae" Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.444009 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd22af6e40a86b54aa68312be29327f0ab663e8f67b7db7c3187a1d823a0c9ae"} err="failed to get container status \"dd22af6e40a86b54aa68312be29327f0ab663e8f67b7db7c3187a1d823a0c9ae\": rpc error: code = NotFound desc = could not find container \"dd22af6e40a86b54aa68312be29327f0ab663e8f67b7db7c3187a1d823a0c9ae\": container with ID starting with dd22af6e40a86b54aa68312be29327f0ab663e8f67b7db7c3187a1d823a0c9ae not found: ID does not exist" Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.444221 4696 scope.go:117] "RemoveContainer" containerID="60d54f52b8f62e3f02ae6a78c86e02e67f31dea89ae92fc551f09f878a24b04a" Mar 21 08:30:56 crc kubenswrapper[4696]: E0321 08:30:56.444784 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60d54f52b8f62e3f02ae6a78c86e02e67f31dea89ae92fc551f09f878a24b04a\": container with ID starting with 60d54f52b8f62e3f02ae6a78c86e02e67f31dea89ae92fc551f09f878a24b04a not found: ID does not exist" containerID="60d54f52b8f62e3f02ae6a78c86e02e67f31dea89ae92fc551f09f878a24b04a" Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.444837 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60d54f52b8f62e3f02ae6a78c86e02e67f31dea89ae92fc551f09f878a24b04a"} err="failed to get container status \"60d54f52b8f62e3f02ae6a78c86e02e67f31dea89ae92fc551f09f878a24b04a\": rpc error: code = NotFound desc = could not find container \"60d54f52b8f62e3f02ae6a78c86e02e67f31dea89ae92fc551f09f878a24b04a\": container with ID starting with 60d54f52b8f62e3f02ae6a78c86e02e67f31dea89ae92fc551f09f878a24b04a not found: ID does not exist" Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.521468 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-trvd2"] Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.521861 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-trvd2" podUID="d6b6d345-41c3-4926-bf33-bac67ac54755" containerName="registry-server" containerID="cri-o://8effc5cfb7634b8f392e55cd369d2a4d9310dce98fa9bd6a01e981736f200fb5" gracePeriod=2 Mar 21 08:30:56 crc kubenswrapper[4696]: I0321 08:30:56.547797 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5149cc5e-d877-4d52-8db6-7df6ae060ab0" path="/var/lib/kubelet/pods/5149cc5e-d877-4d52-8db6-7df6ae060ab0/volumes" Mar 21 08:30:57 crc kubenswrapper[4696]: I0321 08:30:57.257922 4696 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-26 23:03:22.218546202 +0000 UTC Mar 21 08:30:57 crc kubenswrapper[4696]: I0321 08:30:57.258196 4696 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6734h32m24.96035253s for next certificate rotation Mar 21 08:30:58 crc kubenswrapper[4696]: I0321 08:30:58.259164 4696 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-27 20:29:45.778626786 +0000 UTC Mar 21 08:30:58 crc kubenswrapper[4696]: I0321 08:30:58.260503 4696 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6755h58m47.518135307s for next certificate rotation Mar 21 08:30:58 crc kubenswrapper[4696]: I0321 08:30:58.399258 4696 generic.go:334] "Generic (PLEG): container finished" podID="d6b6d345-41c3-4926-bf33-bac67ac54755" containerID="8effc5cfb7634b8f392e55cd369d2a4d9310dce98fa9bd6a01e981736f200fb5" exitCode=0 Mar 21 08:30:58 crc kubenswrapper[4696]: I0321 08:30:58.399319 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trvd2" event={"ID":"d6b6d345-41c3-4926-bf33-bac67ac54755","Type":"ContainerDied","Data":"8effc5cfb7634b8f392e55cd369d2a4d9310dce98fa9bd6a01e981736f200fb5"} Mar 21 08:30:58 crc kubenswrapper[4696]: I0321 08:30:58.400174 4696 generic.go:334] "Generic (PLEG): container finished" podID="a6581056-7677-491f-8f34-5626b43b6e77" containerID="89cca28a71bf3998ea6e1b5417a80f0fb41aadfedf5f4beb2237ac7e4ee7ae9e" exitCode=0 Mar 21 08:30:58 crc kubenswrapper[4696]: I0321 08:30:58.400196 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568030-4x55c" event={"ID":"a6581056-7677-491f-8f34-5626b43b6e77","Type":"ContainerDied","Data":"89cca28a71bf3998ea6e1b5417a80f0fb41aadfedf5f4beb2237ac7e4ee7ae9e"} Mar 21 08:30:58 crc kubenswrapper[4696]: I0321 08:30:58.756028 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-trvd2" Mar 21 08:30:58 crc kubenswrapper[4696]: I0321 08:30:58.813041 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpfnx\" (UniqueName: \"kubernetes.io/projected/d6b6d345-41c3-4926-bf33-bac67ac54755-kube-api-access-dpfnx\") pod \"d6b6d345-41c3-4926-bf33-bac67ac54755\" (UID: \"d6b6d345-41c3-4926-bf33-bac67ac54755\") " Mar 21 08:30:58 crc kubenswrapper[4696]: I0321 08:30:58.813103 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6b6d345-41c3-4926-bf33-bac67ac54755-utilities\") pod \"d6b6d345-41c3-4926-bf33-bac67ac54755\" (UID: \"d6b6d345-41c3-4926-bf33-bac67ac54755\") " Mar 21 08:30:58 crc kubenswrapper[4696]: I0321 08:30:58.813147 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6b6d345-41c3-4926-bf33-bac67ac54755-catalog-content\") pod \"d6b6d345-41c3-4926-bf33-bac67ac54755\" (UID: \"d6b6d345-41c3-4926-bf33-bac67ac54755\") " Mar 21 08:30:58 crc kubenswrapper[4696]: I0321 08:30:58.814135 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6b6d345-41c3-4926-bf33-bac67ac54755-utilities" (OuterVolumeSpecName: "utilities") pod "d6b6d345-41c3-4926-bf33-bac67ac54755" (UID: "d6b6d345-41c3-4926-bf33-bac67ac54755"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:30:58 crc kubenswrapper[4696]: I0321 08:30:58.822132 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6b6d345-41c3-4926-bf33-bac67ac54755-kube-api-access-dpfnx" (OuterVolumeSpecName: "kube-api-access-dpfnx") pod "d6b6d345-41c3-4926-bf33-bac67ac54755" (UID: "d6b6d345-41c3-4926-bf33-bac67ac54755"). InnerVolumeSpecName "kube-api-access-dpfnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:30:58 crc kubenswrapper[4696]: I0321 08:30:58.838911 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6b6d345-41c3-4926-bf33-bac67ac54755-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d6b6d345-41c3-4926-bf33-bac67ac54755" (UID: "d6b6d345-41c3-4926-bf33-bac67ac54755"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:30:58 crc kubenswrapper[4696]: I0321 08:30:58.914737 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpfnx\" (UniqueName: \"kubernetes.io/projected/d6b6d345-41c3-4926-bf33-bac67ac54755-kube-api-access-dpfnx\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:58 crc kubenswrapper[4696]: I0321 08:30:58.914794 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6b6d345-41c3-4926-bf33-bac67ac54755-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:58 crc kubenswrapper[4696]: I0321 08:30:58.914813 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6b6d345-41c3-4926-bf33-bac67ac54755-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 08:30:59 crc kubenswrapper[4696]: I0321 08:30:59.409312 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qsg67" event={"ID":"6aabaca2-7def-4550-bae1-9337b7c1002b","Type":"ContainerStarted","Data":"bdc0c304327cc59729598ff33c42a54e45d127e5d42d5558ef2977e434010ea4"} Mar 21 08:30:59 crc kubenswrapper[4696]: I0321 08:30:59.412178 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trvd2" event={"ID":"d6b6d345-41c3-4926-bf33-bac67ac54755","Type":"ContainerDied","Data":"834509e63dca63e1241ecc534865debc65b471600775671608f17370770b6eb0"} Mar 21 08:30:59 crc kubenswrapper[4696]: I0321 08:30:59.412185 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-trvd2" Mar 21 08:30:59 crc kubenswrapper[4696]: I0321 08:30:59.412281 4696 scope.go:117] "RemoveContainer" containerID="8effc5cfb7634b8f392e55cd369d2a4d9310dce98fa9bd6a01e981736f200fb5" Mar 21 08:30:59 crc kubenswrapper[4696]: I0321 08:30:59.436015 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qsg67" podStartSLOduration=5.226179229 podStartE2EDuration="1m10.435981069s" podCreationTimestamp="2026-03-21 08:29:49 +0000 UTC" firstStartedPulling="2026-03-21 08:29:53.343269775 +0000 UTC m=+127.464150488" lastFinishedPulling="2026-03-21 08:30:58.553071625 +0000 UTC m=+192.673952328" observedRunningTime="2026-03-21 08:30:59.431084629 +0000 UTC m=+193.551965352" watchObservedRunningTime="2026-03-21 08:30:59.435981069 +0000 UTC m=+193.556861822" Mar 21 08:30:59 crc kubenswrapper[4696]: I0321 08:30:59.447111 4696 scope.go:117] "RemoveContainer" containerID="4566194647ac029d7b2db7bbd09a1d36de0d0bec6e2bb6fe2d5ee7b0eba7fc9d" Mar 21 08:30:59 crc kubenswrapper[4696]: I0321 08:30:59.452852 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-trvd2"] Mar 21 08:30:59 crc kubenswrapper[4696]: I0321 08:30:59.456899 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-trvd2"] Mar 21 08:30:59 crc kubenswrapper[4696]: I0321 08:30:59.469351 4696 scope.go:117] "RemoveContainer" containerID="c1faf00a7e367e4b6403ca6d24916fb6f0f207cff786a4717bc6e7d6a0f5c18c" Mar 21 08:30:59 crc kubenswrapper[4696]: I0321 08:30:59.755071 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568030-4x55c" Mar 21 08:30:59 crc kubenswrapper[4696]: I0321 08:30:59.774241 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" podUID="851ce288-d60d-4875-ae01-cc13268fec8c" containerName="oauth-openshift" containerID="cri-o://211d6c3be2967466e12208280fc9b36b1d12586d1058b439a7fc675ea195b4bc" gracePeriod=15 Mar 21 08:30:59 crc kubenswrapper[4696]: I0321 08:30:59.825724 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgqbs\" (UniqueName: \"kubernetes.io/projected/a6581056-7677-491f-8f34-5626b43b6e77-kube-api-access-bgqbs\") pod \"a6581056-7677-491f-8f34-5626b43b6e77\" (UID: \"a6581056-7677-491f-8f34-5626b43b6e77\") " Mar 21 08:30:59 crc kubenswrapper[4696]: I0321 08:30:59.832106 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6581056-7677-491f-8f34-5626b43b6e77-kube-api-access-bgqbs" (OuterVolumeSpecName: "kube-api-access-bgqbs") pod "a6581056-7677-491f-8f34-5626b43b6e77" (UID: "a6581056-7677-491f-8f34-5626b43b6e77"). InnerVolumeSpecName "kube-api-access-bgqbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:30:59 crc kubenswrapper[4696]: I0321 08:30:59.927438 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgqbs\" (UniqueName: \"kubernetes.io/projected/a6581056-7677-491f-8f34-5626b43b6e77-kube-api-access-bgqbs\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:00 crc kubenswrapper[4696]: I0321 08:31:00.420547 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568030-4x55c" event={"ID":"a6581056-7677-491f-8f34-5626b43b6e77","Type":"ContainerDied","Data":"13e173b4a2a99a0f62501dbfef5be4707fb0e0fc4b93c42093c8580e735872e9"} Mar 21 08:31:00 crc kubenswrapper[4696]: I0321 08:31:00.420610 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13e173b4a2a99a0f62501dbfef5be4707fb0e0fc4b93c42093c8580e735872e9" Mar 21 08:31:00 crc kubenswrapper[4696]: I0321 08:31:00.420664 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568030-4x55c" Mar 21 08:31:00 crc kubenswrapper[4696]: I0321 08:31:00.469320 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qsg67" Mar 21 08:31:00 crc kubenswrapper[4696]: I0321 08:31:00.469402 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qsg67" Mar 21 08:31:00 crc kubenswrapper[4696]: I0321 08:31:00.507595 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qsg67" Mar 21 08:31:00 crc kubenswrapper[4696]: I0321 08:31:00.541131 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6b6d345-41c3-4926-bf33-bac67ac54755" path="/var/lib/kubelet/pods/d6b6d345-41c3-4926-bf33-bac67ac54755/volumes" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.432467 4696 generic.go:334] "Generic (PLEG): container finished" podID="851ce288-d60d-4875-ae01-cc13268fec8c" containerID="211d6c3be2967466e12208280fc9b36b1d12586d1058b439a7fc675ea195b4bc" exitCode=0 Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.432968 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" event={"ID":"851ce288-d60d-4875-ae01-cc13268fec8c","Type":"ContainerDied","Data":"211d6c3be2967466e12208280fc9b36b1d12586d1058b439a7fc675ea195b4bc"} Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.490813 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.549384 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-router-certs\") pod \"851ce288-d60d-4875-ae01-cc13268fec8c\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.549440 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6fqx\" (UniqueName: \"kubernetes.io/projected/851ce288-d60d-4875-ae01-cc13268fec8c-kube-api-access-x6fqx\") pod \"851ce288-d60d-4875-ae01-cc13268fec8c\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.549470 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-idp-0-file-data\") pod \"851ce288-d60d-4875-ae01-cc13268fec8c\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.549492 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-service-ca\") pod \"851ce288-d60d-4875-ae01-cc13268fec8c\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.549514 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-cliconfig\") pod \"851ce288-d60d-4875-ae01-cc13268fec8c\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.549536 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-trusted-ca-bundle\") pod \"851ce288-d60d-4875-ae01-cc13268fec8c\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.549581 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-serving-cert\") pod \"851ce288-d60d-4875-ae01-cc13268fec8c\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.549606 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-audit-policies\") pod \"851ce288-d60d-4875-ae01-cc13268fec8c\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.549642 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-ocp-branding-template\") pod \"851ce288-d60d-4875-ae01-cc13268fec8c\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.549676 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-template-provider-selection\") pod \"851ce288-d60d-4875-ae01-cc13268fec8c\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.549713 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-template-error\") pod \"851ce288-d60d-4875-ae01-cc13268fec8c\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.549741 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/851ce288-d60d-4875-ae01-cc13268fec8c-audit-dir\") pod \"851ce288-d60d-4875-ae01-cc13268fec8c\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.549761 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-session\") pod \"851ce288-d60d-4875-ae01-cc13268fec8c\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.549967 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-template-login\") pod \"851ce288-d60d-4875-ae01-cc13268fec8c\" (UID: \"851ce288-d60d-4875-ae01-cc13268fec8c\") " Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.550507 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "851ce288-d60d-4875-ae01-cc13268fec8c" (UID: "851ce288-d60d-4875-ae01-cc13268fec8c"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.550611 4696 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.550873 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "851ce288-d60d-4875-ae01-cc13268fec8c" (UID: "851ce288-d60d-4875-ae01-cc13268fec8c"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.550944 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/851ce288-d60d-4875-ae01-cc13268fec8c-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "851ce288-d60d-4875-ae01-cc13268fec8c" (UID: "851ce288-d60d-4875-ae01-cc13268fec8c"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.551350 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "851ce288-d60d-4875-ae01-cc13268fec8c" (UID: "851ce288-d60d-4875-ae01-cc13268fec8c"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.551493 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "851ce288-d60d-4875-ae01-cc13268fec8c" (UID: "851ce288-d60d-4875-ae01-cc13268fec8c"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.554583 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "851ce288-d60d-4875-ae01-cc13268fec8c" (UID: "851ce288-d60d-4875-ae01-cc13268fec8c"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.555381 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "851ce288-d60d-4875-ae01-cc13268fec8c" (UID: "851ce288-d60d-4875-ae01-cc13268fec8c"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.568980 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "851ce288-d60d-4875-ae01-cc13268fec8c" (UID: "851ce288-d60d-4875-ae01-cc13268fec8c"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.569387 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "851ce288-d60d-4875-ae01-cc13268fec8c" (UID: "851ce288-d60d-4875-ae01-cc13268fec8c"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.569544 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "851ce288-d60d-4875-ae01-cc13268fec8c" (UID: "851ce288-d60d-4875-ae01-cc13268fec8c"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.569632 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/851ce288-d60d-4875-ae01-cc13268fec8c-kube-api-access-x6fqx" (OuterVolumeSpecName: "kube-api-access-x6fqx") pod "851ce288-d60d-4875-ae01-cc13268fec8c" (UID: "851ce288-d60d-4875-ae01-cc13268fec8c"). InnerVolumeSpecName "kube-api-access-x6fqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.579118 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "851ce288-d60d-4875-ae01-cc13268fec8c" (UID: "851ce288-d60d-4875-ae01-cc13268fec8c"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.579369 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "851ce288-d60d-4875-ae01-cc13268fec8c" (UID: "851ce288-d60d-4875-ae01-cc13268fec8c"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.579775 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "851ce288-d60d-4875-ae01-cc13268fec8c" (UID: "851ce288-d60d-4875-ae01-cc13268fec8c"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.652673 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.652717 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6fqx\" (UniqueName: \"kubernetes.io/projected/851ce288-d60d-4875-ae01-cc13268fec8c-kube-api-access-x6fqx\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.652736 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.652753 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.652772 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.652788 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.652804 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.652840 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.652858 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.652877 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.652894 4696 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/851ce288-d60d-4875-ae01-cc13268fec8c-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.652911 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:01 crc kubenswrapper[4696]: I0321 08:31:01.652929 4696 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/851ce288-d60d-4875-ae01-cc13268fec8c-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:02 crc kubenswrapper[4696]: I0321 08:31:02.381708 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cfjs8" Mar 21 08:31:02 crc kubenswrapper[4696]: I0321 08:31:02.426626 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cfjs8" Mar 21 08:31:02 crc kubenswrapper[4696]: I0321 08:31:02.441061 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" event={"ID":"851ce288-d60d-4875-ae01-cc13268fec8c","Type":"ContainerDied","Data":"f63c031a8c919e759aad83e9d49ec2a6b148c4ce571fa9a3c83bc482c55a72f5"} Mar 21 08:31:02 crc kubenswrapper[4696]: I0321 08:31:02.441162 4696 scope.go:117] "RemoveContainer" containerID="211d6c3be2967466e12208280fc9b36b1d12586d1058b439a7fc675ea195b4bc" Mar 21 08:31:02 crc kubenswrapper[4696]: I0321 08:31:02.441948 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-qvm82" Mar 21 08:31:02 crc kubenswrapper[4696]: I0321 08:31:02.491721 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qvm82"] Mar 21 08:31:02 crc kubenswrapper[4696]: I0321 08:31:02.497224 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qvm82"] Mar 21 08:31:02 crc kubenswrapper[4696]: I0321 08:31:02.544467 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="851ce288-d60d-4875-ae01-cc13268fec8c" path="/var/lib/kubelet/pods/851ce288-d60d-4875-ae01-cc13268fec8c/volumes" Mar 21 08:31:02 crc kubenswrapper[4696]: I0321 08:31:02.766801 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jmqr7" Mar 21 08:31:02 crc kubenswrapper[4696]: I0321 08:31:02.811372 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jmqr7" Mar 21 08:31:04 crc kubenswrapper[4696]: I0321 08:31:04.920338 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jmqr7"] Mar 21 08:31:04 crc kubenswrapper[4696]: I0321 08:31:04.921071 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jmqr7" podUID="6ed31bad-6fb5-49be-922b-27198867ca4a" containerName="registry-server" containerID="cri-o://2fb2abd4ababf21c2f49afb05b5050ad78ef3a7087ba8a4a1a5adc52ed8a1362" gracePeriod=2 Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.462618 4696 generic.go:334] "Generic (PLEG): container finished" podID="6ed31bad-6fb5-49be-922b-27198867ca4a" containerID="2fb2abd4ababf21c2f49afb05b5050ad78ef3a7087ba8a4a1a5adc52ed8a1362" exitCode=0 Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.462673 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jmqr7" event={"ID":"6ed31bad-6fb5-49be-922b-27198867ca4a","Type":"ContainerDied","Data":"2fb2abd4ababf21c2f49afb05b5050ad78ef3a7087ba8a4a1a5adc52ed8a1362"} Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.502225 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7dd5d7c474-spgtg"] Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.502888 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" podUID="3ce87a6d-6ef8-4c95-8f74-d1597371ae1a" containerName="controller-manager" containerID="cri-o://0e2c8238b924613669a5ddf1009989ad85ce102a5eb501a92e47f8bac56b0670" gracePeriod=30 Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.592869 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj"] Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.593105 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" podUID="fd34f488-8a79-4e8f-8fdf-baaeefc325ad" containerName="route-controller-manager" containerID="cri-o://8939678ed5699cc486f5bd92aa356cf2e8ce9c133d16a3aa9893a3ba4a8f1c3c" gracePeriod=30 Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.873637 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jmqr7" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.908881 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ed31bad-6fb5-49be-922b-27198867ca4a-catalog-content\") pod \"6ed31bad-6fb5-49be-922b-27198867ca4a\" (UID: \"6ed31bad-6fb5-49be-922b-27198867ca4a\") " Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.908976 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9q4bs\" (UniqueName: \"kubernetes.io/projected/6ed31bad-6fb5-49be-922b-27198867ca4a-kube-api-access-9q4bs\") pod \"6ed31bad-6fb5-49be-922b-27198867ca4a\" (UID: \"6ed31bad-6fb5-49be-922b-27198867ca4a\") " Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.909013 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ed31bad-6fb5-49be-922b-27198867ca4a-utilities\") pod \"6ed31bad-6fb5-49be-922b-27198867ca4a\" (UID: \"6ed31bad-6fb5-49be-922b-27198867ca4a\") " Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.910723 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ed31bad-6fb5-49be-922b-27198867ca4a-utilities" (OuterVolumeSpecName: "utilities") pod "6ed31bad-6fb5-49be-922b-27198867ca4a" (UID: "6ed31bad-6fb5-49be-922b-27198867ca4a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.919604 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ed31bad-6fb5-49be-922b-27198867ca4a-kube-api-access-9q4bs" (OuterVolumeSpecName: "kube-api-access-9q4bs") pod "6ed31bad-6fb5-49be-922b-27198867ca4a" (UID: "6ed31bad-6fb5-49be-922b-27198867ca4a"). InnerVolumeSpecName "kube-api-access-9q4bs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.957407 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5cc5b65bd-xv4st"] Mar 21 08:31:05 crc kubenswrapper[4696]: E0321 08:31:05.957670 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6b6d345-41c3-4926-bf33-bac67ac54755" containerName="registry-server" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.957682 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6b6d345-41c3-4926-bf33-bac67ac54755" containerName="registry-server" Mar 21 08:31:05 crc kubenswrapper[4696]: E0321 08:31:05.957694 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ed31bad-6fb5-49be-922b-27198867ca4a" containerName="extract-content" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.957702 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ed31bad-6fb5-49be-922b-27198867ca4a" containerName="extract-content" Mar 21 08:31:05 crc kubenswrapper[4696]: E0321 08:31:05.957711 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6b6d345-41c3-4926-bf33-bac67ac54755" containerName="extract-utilities" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.957718 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6b6d345-41c3-4926-bf33-bac67ac54755" containerName="extract-utilities" Mar 21 08:31:05 crc kubenswrapper[4696]: E0321 08:31:05.957726 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ed31bad-6fb5-49be-922b-27198867ca4a" containerName="extract-utilities" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.957732 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ed31bad-6fb5-49be-922b-27198867ca4a" containerName="extract-utilities" Mar 21 08:31:05 crc kubenswrapper[4696]: E0321 08:31:05.957745 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5149cc5e-d877-4d52-8db6-7df6ae060ab0" containerName="registry-server" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.957751 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="5149cc5e-d877-4d52-8db6-7df6ae060ab0" containerName="registry-server" Mar 21 08:31:05 crc kubenswrapper[4696]: E0321 08:31:05.957759 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6b6d345-41c3-4926-bf33-bac67ac54755" containerName="extract-content" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.957765 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6b6d345-41c3-4926-bf33-bac67ac54755" containerName="extract-content" Mar 21 08:31:05 crc kubenswrapper[4696]: E0321 08:31:05.957774 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ed31bad-6fb5-49be-922b-27198867ca4a" containerName="registry-server" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.957780 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ed31bad-6fb5-49be-922b-27198867ca4a" containerName="registry-server" Mar 21 08:31:05 crc kubenswrapper[4696]: E0321 08:31:05.957790 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6581056-7677-491f-8f34-5626b43b6e77" containerName="oc" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.957796 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6581056-7677-491f-8f34-5626b43b6e77" containerName="oc" Mar 21 08:31:05 crc kubenswrapper[4696]: E0321 08:31:05.957804 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="851ce288-d60d-4875-ae01-cc13268fec8c" containerName="oauth-openshift" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.957858 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="851ce288-d60d-4875-ae01-cc13268fec8c" containerName="oauth-openshift" Mar 21 08:31:05 crc kubenswrapper[4696]: E0321 08:31:05.957868 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5149cc5e-d877-4d52-8db6-7df6ae060ab0" containerName="extract-content" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.957873 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="5149cc5e-d877-4d52-8db6-7df6ae060ab0" containerName="extract-content" Mar 21 08:31:05 crc kubenswrapper[4696]: E0321 08:31:05.957882 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5149cc5e-d877-4d52-8db6-7df6ae060ab0" containerName="extract-utilities" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.957887 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="5149cc5e-d877-4d52-8db6-7df6ae060ab0" containerName="extract-utilities" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.957988 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ed31bad-6fb5-49be-922b-27198867ca4a" containerName="registry-server" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.957999 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="5149cc5e-d877-4d52-8db6-7df6ae060ab0" containerName="registry-server" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.958009 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6b6d345-41c3-4926-bf33-bac67ac54755" containerName="registry-server" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.958018 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="851ce288-d60d-4875-ae01-cc13268fec8c" containerName="oauth-openshift" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.958027 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6581056-7677-491f-8f34-5626b43b6e77" containerName="oc" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.958557 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.961288 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5cc5b65bd-xv4st"] Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.967354 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.968722 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.968853 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.969060 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.969149 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.969309 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.969590 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.969661 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.969874 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.970184 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.970280 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.970674 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.979267 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.982254 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 21 08:31:05 crc kubenswrapper[4696]: I0321 08:31:05.987513 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.010783 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a1596087-33a6-444c-9352-e6d3375b9340-audit-dir\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.010843 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.010883 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-user-template-login\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.010901 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.010921 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-system-session\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.010940 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.010968 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-system-service-ca\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.010984 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a1596087-33a6-444c-9352-e6d3375b9340-audit-policies\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.011000 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5th6x\" (UniqueName: \"kubernetes.io/projected/a1596087-33a6-444c-9352-e6d3375b9340-kube-api-access-5th6x\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.011017 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.011037 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.011052 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.011071 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-system-router-certs\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.011089 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-user-template-error\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.011122 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9q4bs\" (UniqueName: \"kubernetes.io/projected/6ed31bad-6fb5-49be-922b-27198867ca4a-kube-api-access-9q4bs\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.011133 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ed31bad-6fb5-49be-922b-27198867ca4a-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.067584 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ed31bad-6fb5-49be-922b-27198867ca4a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ed31bad-6fb5-49be-922b-27198867ca4a" (UID: "6ed31bad-6fb5-49be-922b-27198867ca4a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.070965 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.111843 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wp4nj\" (UniqueName: \"kubernetes.io/projected/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-kube-api-access-wp4nj\") pod \"fd34f488-8a79-4e8f-8fdf-baaeefc325ad\" (UID: \"fd34f488-8a79-4e8f-8fdf-baaeefc325ad\") " Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.111977 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-config\") pod \"fd34f488-8a79-4e8f-8fdf-baaeefc325ad\" (UID: \"fd34f488-8a79-4e8f-8fdf-baaeefc325ad\") " Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.112116 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-client-ca\") pod \"fd34f488-8a79-4e8f-8fdf-baaeefc325ad\" (UID: \"fd34f488-8a79-4e8f-8fdf-baaeefc325ad\") " Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.112213 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-serving-cert\") pod \"fd34f488-8a79-4e8f-8fdf-baaeefc325ad\" (UID: \"fd34f488-8a79-4e8f-8fdf-baaeefc325ad\") " Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.112382 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-system-router-certs\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.112889 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-user-template-error\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.113025 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.112689 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-client-ca" (OuterVolumeSpecName: "client-ca") pod "fd34f488-8a79-4e8f-8fdf-baaeefc325ad" (UID: "fd34f488-8a79-4e8f-8fdf-baaeefc325ad"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.112905 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-config" (OuterVolumeSpecName: "config") pod "fd34f488-8a79-4e8f-8fdf-baaeefc325ad" (UID: "fd34f488-8a79-4e8f-8fdf-baaeefc325ad"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.113447 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a1596087-33a6-444c-9352-e6d3375b9340-audit-dir\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.113107 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a1596087-33a6-444c-9352-e6d3375b9340-audit-dir\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.113696 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-user-template-login\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.113800 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.113935 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-system-session\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.114012 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.114103 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-system-service-ca\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.114187 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a1596087-33a6-444c-9352-e6d3375b9340-audit-policies\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.114260 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5th6x\" (UniqueName: \"kubernetes.io/projected/a1596087-33a6-444c-9352-e6d3375b9340-kube-api-access-5th6x\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.114336 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.114409 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.114533 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.114651 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.114721 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ed31bad-6fb5-49be-922b-27198867ca4a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.114788 4696 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.115008 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "fd34f488-8a79-4e8f-8fdf-baaeefc325ad" (UID: "fd34f488-8a79-4e8f-8fdf-baaeefc325ad"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.115683 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-kube-api-access-wp4nj" (OuterVolumeSpecName: "kube-api-access-wp4nj") pod "fd34f488-8a79-4e8f-8fdf-baaeefc325ad" (UID: "fd34f488-8a79-4e8f-8fdf-baaeefc325ad"). InnerVolumeSpecName "kube-api-access-wp4nj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.115732 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-user-template-error\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.116291 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a1596087-33a6-444c-9352-e6d3375b9340-audit-policies\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.116343 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.117032 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-system-service-ca\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.117730 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.118407 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.118622 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.119142 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-user-template-login\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.119585 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.125617 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-system-router-certs\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.128708 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.128918 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a1596087-33a6-444c-9352-e6d3375b9340-v4-0-config-system-session\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.130398 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5th6x\" (UniqueName: \"kubernetes.io/projected/a1596087-33a6-444c-9352-e6d3375b9340-kube-api-access-5th6x\") pod \"oauth-openshift-5cc5b65bd-xv4st\" (UID: \"a1596087-33a6-444c-9352-e6d3375b9340\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.136842 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.215980 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-serving-cert\") pod \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\" (UID: \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\") " Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.216049 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zf2pd\" (UniqueName: \"kubernetes.io/projected/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-kube-api-access-zf2pd\") pod \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\" (UID: \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\") " Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.216111 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-client-ca\") pod \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\" (UID: \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\") " Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.216197 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-config\") pod \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\" (UID: \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\") " Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.216215 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-proxy-ca-bundles\") pod \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\" (UID: \"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a\") " Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.216490 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wp4nj\" (UniqueName: \"kubernetes.io/projected/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-kube-api-access-wp4nj\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.216503 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd34f488-8a79-4e8f-8fdf-baaeefc325ad-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.217247 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "3ce87a6d-6ef8-4c95-8f74-d1597371ae1a" (UID: "3ce87a6d-6ef8-4c95-8f74-d1597371ae1a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.217238 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-client-ca" (OuterVolumeSpecName: "client-ca") pod "3ce87a6d-6ef8-4c95-8f74-d1597371ae1a" (UID: "3ce87a6d-6ef8-4c95-8f74-d1597371ae1a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.217409 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-config" (OuterVolumeSpecName: "config") pod "3ce87a6d-6ef8-4c95-8f74-d1597371ae1a" (UID: "3ce87a6d-6ef8-4c95-8f74-d1597371ae1a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.219401 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3ce87a6d-6ef8-4c95-8f74-d1597371ae1a" (UID: "3ce87a6d-6ef8-4c95-8f74-d1597371ae1a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.219452 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-kube-api-access-zf2pd" (OuterVolumeSpecName: "kube-api-access-zf2pd") pod "3ce87a6d-6ef8-4c95-8f74-d1597371ae1a" (UID: "3ce87a6d-6ef8-4c95-8f74-d1597371ae1a"). InnerVolumeSpecName "kube-api-access-zf2pd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.281788 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.317215 4696 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.317395 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.317471 4696 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.317779 4696 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.317876 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zf2pd\" (UniqueName: \"kubernetes.io/projected/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a-kube-api-access-zf2pd\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.471468 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jmqr7" event={"ID":"6ed31bad-6fb5-49be-922b-27198867ca4a","Type":"ContainerDied","Data":"33de0d1956eb6a34068d548181deedcb4c2b44fcc00a13aeb01ed22f8052fbc9"} Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.471516 4696 scope.go:117] "RemoveContainer" containerID="2fb2abd4ababf21c2f49afb05b5050ad78ef3a7087ba8a4a1a5adc52ed8a1362" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.471622 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jmqr7" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.475333 4696 generic.go:334] "Generic (PLEG): container finished" podID="fd34f488-8a79-4e8f-8fdf-baaeefc325ad" containerID="8939678ed5699cc486f5bd92aa356cf2e8ce9c133d16a3aa9893a3ba4a8f1c3c" exitCode=0 Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.475413 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.475423 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" event={"ID":"fd34f488-8a79-4e8f-8fdf-baaeefc325ad","Type":"ContainerDied","Data":"8939678ed5699cc486f5bd92aa356cf2e8ce9c133d16a3aa9893a3ba4a8f1c3c"} Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.475455 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj" event={"ID":"fd34f488-8a79-4e8f-8fdf-baaeefc325ad","Type":"ContainerDied","Data":"a379de360a98eb26a3ae56623f9803e071cf77d734691aa3e980a6715b4339c9"} Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.478960 4696 generic.go:334] "Generic (PLEG): container finished" podID="3ce87a6d-6ef8-4c95-8f74-d1597371ae1a" containerID="0e2c8238b924613669a5ddf1009989ad85ce102a5eb501a92e47f8bac56b0670" exitCode=0 Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.478995 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" event={"ID":"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a","Type":"ContainerDied","Data":"0e2c8238b924613669a5ddf1009989ad85ce102a5eb501a92e47f8bac56b0670"} Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.479021 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" event={"ID":"3ce87a6d-6ef8-4c95-8f74-d1597371ae1a","Type":"ContainerDied","Data":"86da31bb47d471661ddae8a74649e35d105ad0aed4ba2b92f7bc892cafdd591c"} Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.479075 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7dd5d7c474-spgtg" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.508624 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jmqr7"] Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.509216 4696 scope.go:117] "RemoveContainer" containerID="2984cd8c015b136bc5bcc3860dd8c7f19b0a93e677a874539ef2aed10847c968" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.511373 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jmqr7"] Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.516705 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj"] Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.527565 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b75455d89-xwqgj"] Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.540596 4696 scope.go:117] "RemoveContainer" containerID="023cc09213253c72ec64db5ed3785b171de83e56ee907fcd9c47b1d913d3da7c" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.543542 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ed31bad-6fb5-49be-922b-27198867ca4a" path="/var/lib/kubelet/pods/6ed31bad-6fb5-49be-922b-27198867ca4a/volumes" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.544145 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd34f488-8a79-4e8f-8fdf-baaeefc325ad" path="/var/lib/kubelet/pods/fd34f488-8a79-4e8f-8fdf-baaeefc325ad/volumes" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.544512 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7dd5d7c474-spgtg"] Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.544539 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7dd5d7c474-spgtg"] Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.555202 4696 scope.go:117] "RemoveContainer" containerID="8939678ed5699cc486f5bd92aa356cf2e8ce9c133d16a3aa9893a3ba4a8f1c3c" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.571625 4696 scope.go:117] "RemoveContainer" containerID="8939678ed5699cc486f5bd92aa356cf2e8ce9c133d16a3aa9893a3ba4a8f1c3c" Mar 21 08:31:06 crc kubenswrapper[4696]: E0321 08:31:06.572084 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8939678ed5699cc486f5bd92aa356cf2e8ce9c133d16a3aa9893a3ba4a8f1c3c\": container with ID starting with 8939678ed5699cc486f5bd92aa356cf2e8ce9c133d16a3aa9893a3ba4a8f1c3c not found: ID does not exist" containerID="8939678ed5699cc486f5bd92aa356cf2e8ce9c133d16a3aa9893a3ba4a8f1c3c" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.572133 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8939678ed5699cc486f5bd92aa356cf2e8ce9c133d16a3aa9893a3ba4a8f1c3c"} err="failed to get container status \"8939678ed5699cc486f5bd92aa356cf2e8ce9c133d16a3aa9893a3ba4a8f1c3c\": rpc error: code = NotFound desc = could not find container \"8939678ed5699cc486f5bd92aa356cf2e8ce9c133d16a3aa9893a3ba4a8f1c3c\": container with ID starting with 8939678ed5699cc486f5bd92aa356cf2e8ce9c133d16a3aa9893a3ba4a8f1c3c not found: ID does not exist" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.572161 4696 scope.go:117] "RemoveContainer" containerID="0e2c8238b924613669a5ddf1009989ad85ce102a5eb501a92e47f8bac56b0670" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.585992 4696 scope.go:117] "RemoveContainer" containerID="0e2c8238b924613669a5ddf1009989ad85ce102a5eb501a92e47f8bac56b0670" Mar 21 08:31:06 crc kubenswrapper[4696]: E0321 08:31:06.586327 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e2c8238b924613669a5ddf1009989ad85ce102a5eb501a92e47f8bac56b0670\": container with ID starting with 0e2c8238b924613669a5ddf1009989ad85ce102a5eb501a92e47f8bac56b0670 not found: ID does not exist" containerID="0e2c8238b924613669a5ddf1009989ad85ce102a5eb501a92e47f8bac56b0670" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.586362 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e2c8238b924613669a5ddf1009989ad85ce102a5eb501a92e47f8bac56b0670"} err="failed to get container status \"0e2c8238b924613669a5ddf1009989ad85ce102a5eb501a92e47f8bac56b0670\": rpc error: code = NotFound desc = could not find container \"0e2c8238b924613669a5ddf1009989ad85ce102a5eb501a92e47f8bac56b0670\": container with ID starting with 0e2c8238b924613669a5ddf1009989ad85ce102a5eb501a92e47f8bac56b0670 not found: ID does not exist" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.743380 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5cc5b65bd-xv4st"] Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.943487 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr"] Mar 21 08:31:06 crc kubenswrapper[4696]: E0321 08:31:06.944058 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd34f488-8a79-4e8f-8fdf-baaeefc325ad" containerName="route-controller-manager" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.944418 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd34f488-8a79-4e8f-8fdf-baaeefc325ad" containerName="route-controller-manager" Mar 21 08:31:06 crc kubenswrapper[4696]: E0321 08:31:06.944563 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ce87a6d-6ef8-4c95-8f74-d1597371ae1a" containerName="controller-manager" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.944711 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ce87a6d-6ef8-4c95-8f74-d1597371ae1a" containerName="controller-manager" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.945048 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ce87a6d-6ef8-4c95-8f74-d1597371ae1a" containerName="controller-manager" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.945210 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd34f488-8a79-4e8f-8fdf-baaeefc325ad" containerName="route-controller-manager" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.945941 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5f687966f7-266hn"] Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.946134 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.947201 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.950454 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.950545 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.953609 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.955646 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.956130 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.956153 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.956397 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.957624 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.957775 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.957859 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.958165 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.958205 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.964035 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.964726 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f687966f7-266hn"] Mar 21 08:31:06 crc kubenswrapper[4696]: I0321 08:31:06.971614 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr"] Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.025299 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c119122-2dfa-4e3b-a2d3-31a27af3593f-config\") pod \"route-controller-manager-54644864cb-r4vgr\" (UID: \"3c119122-2dfa-4e3b-a2d3-31a27af3593f\") " pod="openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.025359 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh7hp\" (UniqueName: \"kubernetes.io/projected/3c119122-2dfa-4e3b-a2d3-31a27af3593f-kube-api-access-kh7hp\") pod \"route-controller-manager-54644864cb-r4vgr\" (UID: \"3c119122-2dfa-4e3b-a2d3-31a27af3593f\") " pod="openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.025407 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5edbe45a-6b0a-4a86-9a30-9325c8465d50-proxy-ca-bundles\") pod \"controller-manager-5f687966f7-266hn\" (UID: \"5edbe45a-6b0a-4a86-9a30-9325c8465d50\") " pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.025671 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5edbe45a-6b0a-4a86-9a30-9325c8465d50-config\") pod \"controller-manager-5f687966f7-266hn\" (UID: \"5edbe45a-6b0a-4a86-9a30-9325c8465d50\") " pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.025909 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5edbe45a-6b0a-4a86-9a30-9325c8465d50-serving-cert\") pod \"controller-manager-5f687966f7-266hn\" (UID: \"5edbe45a-6b0a-4a86-9a30-9325c8465d50\") " pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.025973 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wj4f\" (UniqueName: \"kubernetes.io/projected/5edbe45a-6b0a-4a86-9a30-9325c8465d50-kube-api-access-8wj4f\") pod \"controller-manager-5f687966f7-266hn\" (UID: \"5edbe45a-6b0a-4a86-9a30-9325c8465d50\") " pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.026043 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5edbe45a-6b0a-4a86-9a30-9325c8465d50-client-ca\") pod \"controller-manager-5f687966f7-266hn\" (UID: \"5edbe45a-6b0a-4a86-9a30-9325c8465d50\") " pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.026070 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3c119122-2dfa-4e3b-a2d3-31a27af3593f-client-ca\") pod \"route-controller-manager-54644864cb-r4vgr\" (UID: \"3c119122-2dfa-4e3b-a2d3-31a27af3593f\") " pod="openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.026117 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c119122-2dfa-4e3b-a2d3-31a27af3593f-serving-cert\") pod \"route-controller-manager-54644864cb-r4vgr\" (UID: \"3c119122-2dfa-4e3b-a2d3-31a27af3593f\") " pod="openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.127280 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3c119122-2dfa-4e3b-a2d3-31a27af3593f-client-ca\") pod \"route-controller-manager-54644864cb-r4vgr\" (UID: \"3c119122-2dfa-4e3b-a2d3-31a27af3593f\") " pod="openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.127673 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c119122-2dfa-4e3b-a2d3-31a27af3593f-serving-cert\") pod \"route-controller-manager-54644864cb-r4vgr\" (UID: \"3c119122-2dfa-4e3b-a2d3-31a27af3593f\") " pod="openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.127731 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c119122-2dfa-4e3b-a2d3-31a27af3593f-config\") pod \"route-controller-manager-54644864cb-r4vgr\" (UID: \"3c119122-2dfa-4e3b-a2d3-31a27af3593f\") " pod="openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.127770 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh7hp\" (UniqueName: \"kubernetes.io/projected/3c119122-2dfa-4e3b-a2d3-31a27af3593f-kube-api-access-kh7hp\") pod \"route-controller-manager-54644864cb-r4vgr\" (UID: \"3c119122-2dfa-4e3b-a2d3-31a27af3593f\") " pod="openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.127830 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5edbe45a-6b0a-4a86-9a30-9325c8465d50-proxy-ca-bundles\") pod \"controller-manager-5f687966f7-266hn\" (UID: \"5edbe45a-6b0a-4a86-9a30-9325c8465d50\") " pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.127909 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5edbe45a-6b0a-4a86-9a30-9325c8465d50-config\") pod \"controller-manager-5f687966f7-266hn\" (UID: \"5edbe45a-6b0a-4a86-9a30-9325c8465d50\") " pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.127960 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5edbe45a-6b0a-4a86-9a30-9325c8465d50-serving-cert\") pod \"controller-manager-5f687966f7-266hn\" (UID: \"5edbe45a-6b0a-4a86-9a30-9325c8465d50\") " pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.128004 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wj4f\" (UniqueName: \"kubernetes.io/projected/5edbe45a-6b0a-4a86-9a30-9325c8465d50-kube-api-access-8wj4f\") pod \"controller-manager-5f687966f7-266hn\" (UID: \"5edbe45a-6b0a-4a86-9a30-9325c8465d50\") " pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.128058 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5edbe45a-6b0a-4a86-9a30-9325c8465d50-client-ca\") pod \"controller-manager-5f687966f7-266hn\" (UID: \"5edbe45a-6b0a-4a86-9a30-9325c8465d50\") " pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.129526 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5edbe45a-6b0a-4a86-9a30-9325c8465d50-proxy-ca-bundles\") pod \"controller-manager-5f687966f7-266hn\" (UID: \"5edbe45a-6b0a-4a86-9a30-9325c8465d50\") " pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.129629 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5edbe45a-6b0a-4a86-9a30-9325c8465d50-client-ca\") pod \"controller-manager-5f687966f7-266hn\" (UID: \"5edbe45a-6b0a-4a86-9a30-9325c8465d50\") " pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.129914 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5edbe45a-6b0a-4a86-9a30-9325c8465d50-config\") pod \"controller-manager-5f687966f7-266hn\" (UID: \"5edbe45a-6b0a-4a86-9a30-9325c8465d50\") " pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.130239 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c119122-2dfa-4e3b-a2d3-31a27af3593f-config\") pod \"route-controller-manager-54644864cb-r4vgr\" (UID: \"3c119122-2dfa-4e3b-a2d3-31a27af3593f\") " pod="openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.130675 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3c119122-2dfa-4e3b-a2d3-31a27af3593f-client-ca\") pod \"route-controller-manager-54644864cb-r4vgr\" (UID: \"3c119122-2dfa-4e3b-a2d3-31a27af3593f\") " pod="openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.135003 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5edbe45a-6b0a-4a86-9a30-9325c8465d50-serving-cert\") pod \"controller-manager-5f687966f7-266hn\" (UID: \"5edbe45a-6b0a-4a86-9a30-9325c8465d50\") " pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.135092 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c119122-2dfa-4e3b-a2d3-31a27af3593f-serving-cert\") pod \"route-controller-manager-54644864cb-r4vgr\" (UID: \"3c119122-2dfa-4e3b-a2d3-31a27af3593f\") " pod="openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.150483 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wj4f\" (UniqueName: \"kubernetes.io/projected/5edbe45a-6b0a-4a86-9a30-9325c8465d50-kube-api-access-8wj4f\") pod \"controller-manager-5f687966f7-266hn\" (UID: \"5edbe45a-6b0a-4a86-9a30-9325c8465d50\") " pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.157186 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh7hp\" (UniqueName: \"kubernetes.io/projected/3c119122-2dfa-4e3b-a2d3-31a27af3593f-kube-api-access-kh7hp\") pod \"route-controller-manager-54644864cb-r4vgr\" (UID: \"3c119122-2dfa-4e3b-a2d3-31a27af3593f\") " pod="openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.293398 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.308322 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.491238 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" event={"ID":"a1596087-33a6-444c-9352-e6d3375b9340","Type":"ContainerStarted","Data":"dffae883f6ab45a8093de69688c825dc29a4ce50ff8331e1069e2ce4078202b4"} Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.545902 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr"] Mar 21 08:31:07 crc kubenswrapper[4696]: W0321 08:31:07.554109 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c119122_2dfa_4e3b_a2d3_31a27af3593f.slice/crio-7eed7f529fddf2b51b5e74c67fbca18d88900c9a7dc5d1c2621f81a9c6017cb6 WatchSource:0}: Error finding container 7eed7f529fddf2b51b5e74c67fbca18d88900c9a7dc5d1c2621f81a9c6017cb6: Status 404 returned error can't find the container with id 7eed7f529fddf2b51b5e74c67fbca18d88900c9a7dc5d1c2621f81a9c6017cb6 Mar 21 08:31:07 crc kubenswrapper[4696]: I0321 08:31:07.851183 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f687966f7-266hn"] Mar 21 08:31:07 crc kubenswrapper[4696]: W0321 08:31:07.855810 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5edbe45a_6b0a_4a86_9a30_9325c8465d50.slice/crio-0b1b3e672623c1cdc4f9ea3d2a96d3289a02ecbd772d666f683307d8beffa3d3 WatchSource:0}: Error finding container 0b1b3e672623c1cdc4f9ea3d2a96d3289a02ecbd772d666f683307d8beffa3d3: Status 404 returned error can't find the container with id 0b1b3e672623c1cdc4f9ea3d2a96d3289a02ecbd772d666f683307d8beffa3d3 Mar 21 08:31:08 crc kubenswrapper[4696]: I0321 08:31:08.499772 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" event={"ID":"5edbe45a-6b0a-4a86-9a30-9325c8465d50","Type":"ContainerStarted","Data":"b1b910883af65a987d27bbd6d203dcbc0cffd7d26b8e888231400200d6980b81"} Mar 21 08:31:08 crc kubenswrapper[4696]: I0321 08:31:08.499831 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" event={"ID":"5edbe45a-6b0a-4a86-9a30-9325c8465d50","Type":"ContainerStarted","Data":"0b1b3e672623c1cdc4f9ea3d2a96d3289a02ecbd772d666f683307d8beffa3d3"} Mar 21 08:31:08 crc kubenswrapper[4696]: I0321 08:31:08.499985 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" Mar 21 08:31:08 crc kubenswrapper[4696]: I0321 08:31:08.501272 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" event={"ID":"a1596087-33a6-444c-9352-e6d3375b9340","Type":"ContainerStarted","Data":"03d5b880eb333f5d586b4a9520fa808d8890b96df160431f8a4ebade43189d9d"} Mar 21 08:31:08 crc kubenswrapper[4696]: I0321 08:31:08.501479 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:08 crc kubenswrapper[4696]: I0321 08:31:08.502572 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr" event={"ID":"3c119122-2dfa-4e3b-a2d3-31a27af3593f","Type":"ContainerStarted","Data":"7bfa50cd136f31390fddbc74a0342f2eb1300b060f157e79e2916bb32073b23f"} Mar 21 08:31:08 crc kubenswrapper[4696]: I0321 08:31:08.502602 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr" event={"ID":"3c119122-2dfa-4e3b-a2d3-31a27af3593f","Type":"ContainerStarted","Data":"7eed7f529fddf2b51b5e74c67fbca18d88900c9a7dc5d1c2621f81a9c6017cb6"} Mar 21 08:31:08 crc kubenswrapper[4696]: I0321 08:31:08.502998 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr" Mar 21 08:31:08 crc kubenswrapper[4696]: I0321 08:31:08.506328 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" Mar 21 08:31:08 crc kubenswrapper[4696]: I0321 08:31:08.508717 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" Mar 21 08:31:08 crc kubenswrapper[4696]: I0321 08:31:08.509787 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr" Mar 21 08:31:08 crc kubenswrapper[4696]: I0321 08:31:08.530896 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5f687966f7-266hn" podStartSLOduration=3.530878893 podStartE2EDuration="3.530878893s" podCreationTimestamp="2026-03-21 08:31:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:31:08.526292291 +0000 UTC m=+202.647173004" watchObservedRunningTime="2026-03-21 08:31:08.530878893 +0000 UTC m=+202.651759606" Mar 21 08:31:08 crc kubenswrapper[4696]: I0321 08:31:08.541395 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ce87a6d-6ef8-4c95-8f74-d1597371ae1a" path="/var/lib/kubelet/pods/3ce87a6d-6ef8-4c95-8f74-d1597371ae1a/volumes" Mar 21 08:31:08 crc kubenswrapper[4696]: I0321 08:31:08.559435 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-54644864cb-r4vgr" podStartSLOduration=3.559421614 podStartE2EDuration="3.559421614s" podCreationTimestamp="2026-03-21 08:31:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:31:08.557170235 +0000 UTC m=+202.678050948" watchObservedRunningTime="2026-03-21 08:31:08.559421614 +0000 UTC m=+202.680302327" Mar 21 08:31:08 crc kubenswrapper[4696]: I0321 08:31:08.583104 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5cc5b65bd-xv4st" podStartSLOduration=34.583083505 podStartE2EDuration="34.583083505s" podCreationTimestamp="2026-03-21 08:30:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:31:08.58175032 +0000 UTC m=+202.702631033" watchObservedRunningTime="2026-03-21 08:31:08.583083505 +0000 UTC m=+202.703964218" Mar 21 08:31:10 crc kubenswrapper[4696]: I0321 08:31:10.509000 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qsg67" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.084677 4696 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.085528 4696 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.085658 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.085781 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://f72b31e3f8525ae3cbfa7a2b4ccbb88d4aac21c4e1e802d2cde33aa4408e5b24" gracePeriod=15 Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.085820 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://815425653c2afb43bac59c2ba2c04bfe71edde9aa3372e516741e02a9745bccd" gracePeriod=15 Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.085916 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://824fb0d35d4bdbe666975fd095c6299d742b56dc6c34b2ac87cffd9708c821aa" gracePeriod=15 Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.085958 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://ac98b604eeaf478083bc6291ecded1c1c68b1bb7a844b281c66447fb9b7ff248" gracePeriod=15 Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.085994 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://e83144c8a489a404f9916385b8216651b6fcfa968e843b43b62901adbaab80f1" gracePeriod=15 Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.086331 4696 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 21 08:31:12 crc kubenswrapper[4696]: E0321 08:31:12.086577 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.086588 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 08:31:12 crc kubenswrapper[4696]: E0321 08:31:12.086600 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.086606 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 21 08:31:12 crc kubenswrapper[4696]: E0321 08:31:12.086612 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.086618 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 08:31:12 crc kubenswrapper[4696]: E0321 08:31:12.086627 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.086633 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 21 08:31:12 crc kubenswrapper[4696]: E0321 08:31:12.086642 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.086649 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 21 08:31:12 crc kubenswrapper[4696]: E0321 08:31:12.086658 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.086663 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 08:31:12 crc kubenswrapper[4696]: E0321 08:31:12.086673 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.086679 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 08:31:12 crc kubenswrapper[4696]: E0321 08:31:12.086685 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.086691 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 21 08:31:12 crc kubenswrapper[4696]: E0321 08:31:12.086700 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.086707 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.086796 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.086807 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.086819 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.086838 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.086848 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.086859 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.086868 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 21 08:31:12 crc kubenswrapper[4696]: E0321 08:31:12.086979 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.086988 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.087080 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.087265 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.146659 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.200941 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.201055 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.201082 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.201096 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.201123 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.201142 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.201242 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.201288 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.302415 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.302467 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.302498 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.302534 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.302553 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.302567 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.302573 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.302597 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.302617 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.302607 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.302643 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.302647 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.302658 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.302593 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.302761 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.302883 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.439540 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 08:31:12 crc kubenswrapper[4696]: W0321 08:31:12.463624 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-8065f484ebdef71f85c4dab99606c4fd3f4a4471bdcd71e6f879943359e8c880 WatchSource:0}: Error finding container 8065f484ebdef71f85c4dab99606c4fd3f4a4471bdcd71e6f879943359e8c880: Status 404 returned error can't find the container with id 8065f484ebdef71f85c4dab99606c4fd3f4a4471bdcd71e6f879943359e8c880 Mar 21 08:31:12 crc kubenswrapper[4696]: E0321 08:31:12.468162 4696 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.174:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189ece1042d86f8c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:31:12.466833292 +0000 UTC m=+206.587714005,LastTimestamp:2026-03-21 08:31:12.466833292 +0000 UTC m=+206.587714005,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.540409 4696 generic.go:334] "Generic (PLEG): container finished" podID="df078e35-cb4d-4be7-beaf-7fd6fc463a0c" containerID="fe1f72bf9ef983d3676ebee521a1f3b61709c61c083f362ba4402127ced1fb2f" exitCode=0 Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.544001 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"df078e35-cb4d-4be7-beaf-7fd6fc463a0c","Type":"ContainerDied","Data":"fe1f72bf9ef983d3676ebee521a1f3b61709c61c083f362ba4402127ced1fb2f"} Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.544946 4696 status_manager.go:851] "Failed to get status for pod" podUID="df078e35-cb4d-4be7-beaf-7fd6fc463a0c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.545285 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.547207 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.548611 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.550234 4696 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="815425653c2afb43bac59c2ba2c04bfe71edde9aa3372e516741e02a9745bccd" exitCode=0 Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.550261 4696 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="824fb0d35d4bdbe666975fd095c6299d742b56dc6c34b2ac87cffd9708c821aa" exitCode=0 Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.550273 4696 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ac98b604eeaf478083bc6291ecded1c1c68b1bb7a844b281c66447fb9b7ff248" exitCode=0 Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.550287 4696 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e83144c8a489a404f9916385b8216651b6fcfa968e843b43b62901adbaab80f1" exitCode=2 Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.550356 4696 scope.go:117] "RemoveContainer" containerID="a78fedb8a70e8b5cec10a972e63964e900746253f6add19cbedf2e47af762244" Mar 21 08:31:12 crc kubenswrapper[4696]: I0321 08:31:12.552915 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"8065f484ebdef71f85c4dab99606c4fd3f4a4471bdcd71e6f879943359e8c880"} Mar 21 08:31:13 crc kubenswrapper[4696]: I0321 08:31:13.562158 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 21 08:31:13 crc kubenswrapper[4696]: I0321 08:31:13.565943 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"d5becd6a47f3d578519a657f70c3a6a05c41c399caeb6841448da9b2df4fac5c"} Mar 21 08:31:13 crc kubenswrapper[4696]: I0321 08:31:13.566667 4696 status_manager.go:851] "Failed to get status for pod" podUID="df078e35-cb4d-4be7-beaf-7fd6fc463a0c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:13 crc kubenswrapper[4696]: I0321 08:31:13.567188 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.030360 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.030887 4696 status_manager.go:851] "Failed to get status for pod" podUID="df078e35-cb4d-4be7-beaf-7fd6fc463a0c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.031185 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.046704 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/df078e35-cb4d-4be7-beaf-7fd6fc463a0c-kubelet-dir\") pod \"df078e35-cb4d-4be7-beaf-7fd6fc463a0c\" (UID: \"df078e35-cb4d-4be7-beaf-7fd6fc463a0c\") " Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.046766 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/df078e35-cb4d-4be7-beaf-7fd6fc463a0c-var-lock\") pod \"df078e35-cb4d-4be7-beaf-7fd6fc463a0c\" (UID: \"df078e35-cb4d-4be7-beaf-7fd6fc463a0c\") " Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.046796 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/df078e35-cb4d-4be7-beaf-7fd6fc463a0c-kube-api-access\") pod \"df078e35-cb4d-4be7-beaf-7fd6fc463a0c\" (UID: \"df078e35-cb4d-4be7-beaf-7fd6fc463a0c\") " Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.046945 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df078e35-cb4d-4be7-beaf-7fd6fc463a0c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "df078e35-cb4d-4be7-beaf-7fd6fc463a0c" (UID: "df078e35-cb4d-4be7-beaf-7fd6fc463a0c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.047016 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df078e35-cb4d-4be7-beaf-7fd6fc463a0c-var-lock" (OuterVolumeSpecName: "var-lock") pod "df078e35-cb4d-4be7-beaf-7fd6fc463a0c" (UID: "df078e35-cb4d-4be7-beaf-7fd6fc463a0c"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.052764 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df078e35-cb4d-4be7-beaf-7fd6fc463a0c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "df078e35-cb4d-4be7-beaf-7fd6fc463a0c" (UID: "df078e35-cb4d-4be7-beaf-7fd6fc463a0c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.148912 4696 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/df078e35-cb4d-4be7-beaf-7fd6fc463a0c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.149117 4696 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/df078e35-cb4d-4be7-beaf-7fd6fc463a0c-var-lock\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.149151 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/df078e35-cb4d-4be7-beaf-7fd6fc463a0c-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.447607 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.448949 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.449552 4696 status_manager.go:851] "Failed to get status for pod" podUID="df078e35-cb4d-4be7-beaf-7fd6fc463a0c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.450138 4696 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.450560 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.553325 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.553454 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.553485 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.553584 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.553611 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.553762 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.554232 4696 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.554254 4696 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.554265 4696 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.572182 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"df078e35-cb4d-4be7-beaf-7fd6fc463a0c","Type":"ContainerDied","Data":"71115f4b3ed137978a13e46957c8c7f2ce7791599440e820b2edbda857c2b65b"} Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.572227 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.572237 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71115f4b3ed137978a13e46957c8c7f2ce7791599440e820b2edbda857c2b65b" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.574978 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.576321 4696 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f72b31e3f8525ae3cbfa7a2b4ccbb88d4aac21c4e1e802d2cde33aa4408e5b24" exitCode=0 Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.576479 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.576532 4696 scope.go:117] "RemoveContainer" containerID="815425653c2afb43bac59c2ba2c04bfe71edde9aa3372e516741e02a9745bccd" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.577594 4696 status_manager.go:851] "Failed to get status for pod" podUID="df078e35-cb4d-4be7-beaf-7fd6fc463a0c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.577844 4696 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.578074 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.578294 4696 status_manager.go:851] "Failed to get status for pod" podUID="df078e35-cb4d-4be7-beaf-7fd6fc463a0c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.578548 4696 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.578757 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.593191 4696 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.593636 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.593796 4696 status_manager.go:851] "Failed to get status for pod" podUID="df078e35-cb4d-4be7-beaf-7fd6fc463a0c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.599322 4696 scope.go:117] "RemoveContainer" containerID="824fb0d35d4bdbe666975fd095c6299d742b56dc6c34b2ac87cffd9708c821aa" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.620302 4696 scope.go:117] "RemoveContainer" containerID="ac98b604eeaf478083bc6291ecded1c1c68b1bb7a844b281c66447fb9b7ff248" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.631801 4696 scope.go:117] "RemoveContainer" containerID="e83144c8a489a404f9916385b8216651b6fcfa968e843b43b62901adbaab80f1" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.643314 4696 scope.go:117] "RemoveContainer" containerID="f72b31e3f8525ae3cbfa7a2b4ccbb88d4aac21c4e1e802d2cde33aa4408e5b24" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.659669 4696 scope.go:117] "RemoveContainer" containerID="fede52f3fbd397eb207a4d0daa4587e1987bcd31c175d212557666a4e939701f" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.677187 4696 scope.go:117] "RemoveContainer" containerID="815425653c2afb43bac59c2ba2c04bfe71edde9aa3372e516741e02a9745bccd" Mar 21 08:31:14 crc kubenswrapper[4696]: E0321 08:31:14.678561 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"815425653c2afb43bac59c2ba2c04bfe71edde9aa3372e516741e02a9745bccd\": container with ID starting with 815425653c2afb43bac59c2ba2c04bfe71edde9aa3372e516741e02a9745bccd not found: ID does not exist" containerID="815425653c2afb43bac59c2ba2c04bfe71edde9aa3372e516741e02a9745bccd" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.678650 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"815425653c2afb43bac59c2ba2c04bfe71edde9aa3372e516741e02a9745bccd"} err="failed to get container status \"815425653c2afb43bac59c2ba2c04bfe71edde9aa3372e516741e02a9745bccd\": rpc error: code = NotFound desc = could not find container \"815425653c2afb43bac59c2ba2c04bfe71edde9aa3372e516741e02a9745bccd\": container with ID starting with 815425653c2afb43bac59c2ba2c04bfe71edde9aa3372e516741e02a9745bccd not found: ID does not exist" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.678693 4696 scope.go:117] "RemoveContainer" containerID="824fb0d35d4bdbe666975fd095c6299d742b56dc6c34b2ac87cffd9708c821aa" Mar 21 08:31:14 crc kubenswrapper[4696]: E0321 08:31:14.679351 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"824fb0d35d4bdbe666975fd095c6299d742b56dc6c34b2ac87cffd9708c821aa\": container with ID starting with 824fb0d35d4bdbe666975fd095c6299d742b56dc6c34b2ac87cffd9708c821aa not found: ID does not exist" containerID="824fb0d35d4bdbe666975fd095c6299d742b56dc6c34b2ac87cffd9708c821aa" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.679399 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"824fb0d35d4bdbe666975fd095c6299d742b56dc6c34b2ac87cffd9708c821aa"} err="failed to get container status \"824fb0d35d4bdbe666975fd095c6299d742b56dc6c34b2ac87cffd9708c821aa\": rpc error: code = NotFound desc = could not find container \"824fb0d35d4bdbe666975fd095c6299d742b56dc6c34b2ac87cffd9708c821aa\": container with ID starting with 824fb0d35d4bdbe666975fd095c6299d742b56dc6c34b2ac87cffd9708c821aa not found: ID does not exist" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.679430 4696 scope.go:117] "RemoveContainer" containerID="ac98b604eeaf478083bc6291ecded1c1c68b1bb7a844b281c66447fb9b7ff248" Mar 21 08:31:14 crc kubenswrapper[4696]: E0321 08:31:14.679940 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac98b604eeaf478083bc6291ecded1c1c68b1bb7a844b281c66447fb9b7ff248\": container with ID starting with ac98b604eeaf478083bc6291ecded1c1c68b1bb7a844b281c66447fb9b7ff248 not found: ID does not exist" containerID="ac98b604eeaf478083bc6291ecded1c1c68b1bb7a844b281c66447fb9b7ff248" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.679980 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac98b604eeaf478083bc6291ecded1c1c68b1bb7a844b281c66447fb9b7ff248"} err="failed to get container status \"ac98b604eeaf478083bc6291ecded1c1c68b1bb7a844b281c66447fb9b7ff248\": rpc error: code = NotFound desc = could not find container \"ac98b604eeaf478083bc6291ecded1c1c68b1bb7a844b281c66447fb9b7ff248\": container with ID starting with ac98b604eeaf478083bc6291ecded1c1c68b1bb7a844b281c66447fb9b7ff248 not found: ID does not exist" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.680004 4696 scope.go:117] "RemoveContainer" containerID="e83144c8a489a404f9916385b8216651b6fcfa968e843b43b62901adbaab80f1" Mar 21 08:31:14 crc kubenswrapper[4696]: E0321 08:31:14.680279 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e83144c8a489a404f9916385b8216651b6fcfa968e843b43b62901adbaab80f1\": container with ID starting with e83144c8a489a404f9916385b8216651b6fcfa968e843b43b62901adbaab80f1 not found: ID does not exist" containerID="e83144c8a489a404f9916385b8216651b6fcfa968e843b43b62901adbaab80f1" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.680327 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e83144c8a489a404f9916385b8216651b6fcfa968e843b43b62901adbaab80f1"} err="failed to get container status \"e83144c8a489a404f9916385b8216651b6fcfa968e843b43b62901adbaab80f1\": rpc error: code = NotFound desc = could not find container \"e83144c8a489a404f9916385b8216651b6fcfa968e843b43b62901adbaab80f1\": container with ID starting with e83144c8a489a404f9916385b8216651b6fcfa968e843b43b62901adbaab80f1 not found: ID does not exist" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.680355 4696 scope.go:117] "RemoveContainer" containerID="f72b31e3f8525ae3cbfa7a2b4ccbb88d4aac21c4e1e802d2cde33aa4408e5b24" Mar 21 08:31:14 crc kubenswrapper[4696]: E0321 08:31:14.680630 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f72b31e3f8525ae3cbfa7a2b4ccbb88d4aac21c4e1e802d2cde33aa4408e5b24\": container with ID starting with f72b31e3f8525ae3cbfa7a2b4ccbb88d4aac21c4e1e802d2cde33aa4408e5b24 not found: ID does not exist" containerID="f72b31e3f8525ae3cbfa7a2b4ccbb88d4aac21c4e1e802d2cde33aa4408e5b24" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.680680 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f72b31e3f8525ae3cbfa7a2b4ccbb88d4aac21c4e1e802d2cde33aa4408e5b24"} err="failed to get container status \"f72b31e3f8525ae3cbfa7a2b4ccbb88d4aac21c4e1e802d2cde33aa4408e5b24\": rpc error: code = NotFound desc = could not find container \"f72b31e3f8525ae3cbfa7a2b4ccbb88d4aac21c4e1e802d2cde33aa4408e5b24\": container with ID starting with f72b31e3f8525ae3cbfa7a2b4ccbb88d4aac21c4e1e802d2cde33aa4408e5b24 not found: ID does not exist" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.680712 4696 scope.go:117] "RemoveContainer" containerID="fede52f3fbd397eb207a4d0daa4587e1987bcd31c175d212557666a4e939701f" Mar 21 08:31:14 crc kubenswrapper[4696]: E0321 08:31:14.681002 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fede52f3fbd397eb207a4d0daa4587e1987bcd31c175d212557666a4e939701f\": container with ID starting with fede52f3fbd397eb207a4d0daa4587e1987bcd31c175d212557666a4e939701f not found: ID does not exist" containerID="fede52f3fbd397eb207a4d0daa4587e1987bcd31c175d212557666a4e939701f" Mar 21 08:31:14 crc kubenswrapper[4696]: I0321 08:31:14.681051 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fede52f3fbd397eb207a4d0daa4587e1987bcd31c175d212557666a4e939701f"} err="failed to get container status \"fede52f3fbd397eb207a4d0daa4587e1987bcd31c175d212557666a4e939701f\": rpc error: code = NotFound desc = could not find container \"fede52f3fbd397eb207a4d0daa4587e1987bcd31c175d212557666a4e939701f\": container with ID starting with fede52f3fbd397eb207a4d0daa4587e1987bcd31c175d212557666a4e939701f not found: ID does not exist" Mar 21 08:31:16 crc kubenswrapper[4696]: I0321 08:31:16.538473 4696 status_manager.go:851] "Failed to get status for pod" podUID="df078e35-cb4d-4be7-beaf-7fd6fc463a0c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:16 crc kubenswrapper[4696]: I0321 08:31:16.539084 4696 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:16 crc kubenswrapper[4696]: I0321 08:31:16.539460 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:16 crc kubenswrapper[4696]: I0321 08:31:16.541484 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 21 08:31:17 crc kubenswrapper[4696]: E0321 08:31:17.758493 4696 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.174:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189ece1042d86f8c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 08:31:12.466833292 +0000 UTC m=+206.587714005,LastTimestamp:2026-03-21 08:31:12.466833292 +0000 UTC m=+206.587714005,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 08:31:19 crc kubenswrapper[4696]: E0321 08:31:19.449539 4696 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:19 crc kubenswrapper[4696]: E0321 08:31:19.450204 4696 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:19 crc kubenswrapper[4696]: E0321 08:31:19.450708 4696 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:19 crc kubenswrapper[4696]: E0321 08:31:19.451073 4696 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:19 crc kubenswrapper[4696]: E0321 08:31:19.451587 4696 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:19 crc kubenswrapper[4696]: I0321 08:31:19.451667 4696 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 21 08:31:19 crc kubenswrapper[4696]: E0321 08:31:19.452384 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="200ms" Mar 21 08:31:19 crc kubenswrapper[4696]: E0321 08:31:19.596747 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:31:19Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:31:19Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:31:19Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T08:31:19Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:19 crc kubenswrapper[4696]: E0321 08:31:19.597449 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:19 crc kubenswrapper[4696]: E0321 08:31:19.598143 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:19 crc kubenswrapper[4696]: E0321 08:31:19.598609 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:19 crc kubenswrapper[4696]: E0321 08:31:19.599073 4696 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:19 crc kubenswrapper[4696]: E0321 08:31:19.599110 4696 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 21 08:31:19 crc kubenswrapper[4696]: E0321 08:31:19.653975 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="400ms" Mar 21 08:31:20 crc kubenswrapper[4696]: E0321 08:31:20.055544 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="800ms" Mar 21 08:31:20 crc kubenswrapper[4696]: E0321 08:31:20.548226 4696 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" volumeName="registry-storage" Mar 21 08:31:20 crc kubenswrapper[4696]: E0321 08:31:20.857117 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="1.6s" Mar 21 08:31:22 crc kubenswrapper[4696]: E0321 08:31:22.458805 4696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="3.2s" Mar 21 08:31:23 crc kubenswrapper[4696]: I0321 08:31:23.534141 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:23 crc kubenswrapper[4696]: I0321 08:31:23.534776 4696 status_manager.go:851] "Failed to get status for pod" podUID="df078e35-cb4d-4be7-beaf-7fd6fc463a0c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:23 crc kubenswrapper[4696]: I0321 08:31:23.535362 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:23 crc kubenswrapper[4696]: I0321 08:31:23.547114 4696 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ec2fd6b5-b161-4045-a594-94cd018894f5" Mar 21 08:31:23 crc kubenswrapper[4696]: I0321 08:31:23.547153 4696 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ec2fd6b5-b161-4045-a594-94cd018894f5" Mar 21 08:31:23 crc kubenswrapper[4696]: E0321 08:31:23.547783 4696 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:23 crc kubenswrapper[4696]: I0321 08:31:23.549003 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:23 crc kubenswrapper[4696]: I0321 08:31:23.652199 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6877ebeda17d40166e64dba4cc625ab5b43f6b24f4db16d5646d408f5f6dcf30"} Mar 21 08:31:24 crc kubenswrapper[4696]: I0321 08:31:24.661062 4696 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="1a5b0534f1eb1957f68d455a679dc35835008526df576fd8fc3618e92341ba06" exitCode=0 Mar 21 08:31:24 crc kubenswrapper[4696]: I0321 08:31:24.661124 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"1a5b0534f1eb1957f68d455a679dc35835008526df576fd8fc3618e92341ba06"} Mar 21 08:31:24 crc kubenswrapper[4696]: I0321 08:31:24.661589 4696 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ec2fd6b5-b161-4045-a594-94cd018894f5" Mar 21 08:31:24 crc kubenswrapper[4696]: I0321 08:31:24.661626 4696 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ec2fd6b5-b161-4045-a594-94cd018894f5" Mar 21 08:31:24 crc kubenswrapper[4696]: I0321 08:31:24.662045 4696 status_manager.go:851] "Failed to get status for pod" podUID="df078e35-cb4d-4be7-beaf-7fd6fc463a0c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:24 crc kubenswrapper[4696]: E0321 08:31:24.662265 4696 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:24 crc kubenswrapper[4696]: I0321 08:31:24.662410 4696 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 21 08:31:25 crc kubenswrapper[4696]: I0321 08:31:25.683764 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"26eb532f9f9f916b80f4bbe180e7ca7bdd3b271e46d156a54a6bd1b7e22897cb"} Mar 21 08:31:25 crc kubenswrapper[4696]: I0321 08:31:25.684109 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a2a02efc6fffdeb718db710482b69b747da4d64a251f50f628980443f0df4a56"} Mar 21 08:31:25 crc kubenswrapper[4696]: I0321 08:31:25.684137 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"851c69d3592745a113fe612e43a481921840d096ee1290f514b87ff2aae27f86"} Mar 21 08:31:25 crc kubenswrapper[4696]: I0321 08:31:25.684150 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e99e6074d56dcb34e580638f6e87f993936ff7003170f61dbb968619dd85313d"} Mar 21 08:31:26 crc kubenswrapper[4696]: I0321 08:31:26.692133 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"19c55e17946ededcb8eb6081c09cdbc3bf6e9d3ab1c59bf4409c705a8a6e5496"} Mar 21 08:31:26 crc kubenswrapper[4696]: I0321 08:31:26.692332 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:26 crc kubenswrapper[4696]: I0321 08:31:26.692366 4696 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ec2fd6b5-b161-4045-a594-94cd018894f5" Mar 21 08:31:26 crc kubenswrapper[4696]: I0321 08:31:26.692390 4696 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ec2fd6b5-b161-4045-a594-94cd018894f5" Mar 21 08:31:26 crc kubenswrapper[4696]: I0321 08:31:26.695380 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 21 08:31:26 crc kubenswrapper[4696]: I0321 08:31:26.695884 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 21 08:31:26 crc kubenswrapper[4696]: I0321 08:31:26.695932 4696 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="f4111d26588a9cb6e28559a10dcc9a940428b646a3072500c455e099ab8eb5ee" exitCode=1 Mar 21 08:31:26 crc kubenswrapper[4696]: I0321 08:31:26.695965 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"f4111d26588a9cb6e28559a10dcc9a940428b646a3072500c455e099ab8eb5ee"} Mar 21 08:31:26 crc kubenswrapper[4696]: I0321 08:31:26.696492 4696 scope.go:117] "RemoveContainer" containerID="f4111d26588a9cb6e28559a10dcc9a940428b646a3072500c455e099ab8eb5ee" Mar 21 08:31:27 crc kubenswrapper[4696]: I0321 08:31:27.704221 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 21 08:31:27 crc kubenswrapper[4696]: I0321 08:31:27.705161 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 21 08:31:27 crc kubenswrapper[4696]: I0321 08:31:27.705207 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3f3a987505d69c5be055bce89175fbb2c88a9ff8e546be20b938c00678bfb8d0"} Mar 21 08:31:28 crc kubenswrapper[4696]: I0321 08:31:28.550086 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:28 crc kubenswrapper[4696]: I0321 08:31:28.550147 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:28 crc kubenswrapper[4696]: I0321 08:31:28.556625 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:30 crc kubenswrapper[4696]: I0321 08:31:30.341323 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:31:30 crc kubenswrapper[4696]: I0321 08:31:30.341914 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:31:31 crc kubenswrapper[4696]: I0321 08:31:31.700171 4696 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:31 crc kubenswrapper[4696]: I0321 08:31:31.726248 4696 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ec2fd6b5-b161-4045-a594-94cd018894f5" Mar 21 08:31:31 crc kubenswrapper[4696]: I0321 08:31:31.726270 4696 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ec2fd6b5-b161-4045-a594-94cd018894f5" Mar 21 08:31:31 crc kubenswrapper[4696]: I0321 08:31:31.729714 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:31 crc kubenswrapper[4696]: I0321 08:31:31.733497 4696 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="cd4a71f9-5566-40b3-aa1b-812d2c77b542" Mar 21 08:31:32 crc kubenswrapper[4696]: I0321 08:31:32.101310 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:31:32 crc kubenswrapper[4696]: I0321 08:31:32.106991 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:31:32 crc kubenswrapper[4696]: I0321 08:31:32.730593 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:31:32 crc kubenswrapper[4696]: I0321 08:31:32.730621 4696 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ec2fd6b5-b161-4045-a594-94cd018894f5" Mar 21 08:31:32 crc kubenswrapper[4696]: I0321 08:31:32.730644 4696 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ec2fd6b5-b161-4045-a594-94cd018894f5" Mar 21 08:31:36 crc kubenswrapper[4696]: I0321 08:31:36.557065 4696 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="cd4a71f9-5566-40b3-aa1b-812d2c77b542" Mar 21 08:31:41 crc kubenswrapper[4696]: I0321 08:31:41.434123 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 21 08:31:42 crc kubenswrapper[4696]: I0321 08:31:42.993721 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 21 08:31:43 crc kubenswrapper[4696]: I0321 08:31:43.469708 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 21 08:31:43 crc kubenswrapper[4696]: I0321 08:31:43.598109 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 21 08:31:43 crc kubenswrapper[4696]: I0321 08:31:43.788208 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 21 08:31:43 crc kubenswrapper[4696]: I0321 08:31:43.874499 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 21 08:31:44 crc kubenswrapper[4696]: I0321 08:31:44.216478 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 21 08:31:44 crc kubenswrapper[4696]: I0321 08:31:44.319101 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 21 08:31:44 crc kubenswrapper[4696]: I0321 08:31:44.373072 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 21 08:31:44 crc kubenswrapper[4696]: I0321 08:31:44.389589 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 21 08:31:44 crc kubenswrapper[4696]: I0321 08:31:44.729053 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 21 08:31:44 crc kubenswrapper[4696]: I0321 08:31:44.980160 4696 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 21 08:31:45 crc kubenswrapper[4696]: I0321 08:31:45.153075 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 21 08:31:45 crc kubenswrapper[4696]: I0321 08:31:45.156634 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 08:31:45 crc kubenswrapper[4696]: I0321 08:31:45.175106 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 21 08:31:45 crc kubenswrapper[4696]: I0321 08:31:45.505712 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 21 08:31:45 crc kubenswrapper[4696]: I0321 08:31:45.532795 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 21 08:31:45 crc kubenswrapper[4696]: I0321 08:31:45.551706 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 21 08:31:45 crc kubenswrapper[4696]: I0321 08:31:45.602575 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 21 08:31:45 crc kubenswrapper[4696]: I0321 08:31:45.741720 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 21 08:31:45 crc kubenswrapper[4696]: I0321 08:31:45.753166 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 21 08:31:45 crc kubenswrapper[4696]: I0321 08:31:45.797069 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 21 08:31:45 crc kubenswrapper[4696]: I0321 08:31:45.971367 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 21 08:31:46 crc kubenswrapper[4696]: I0321 08:31:46.018969 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 21 08:31:46 crc kubenswrapper[4696]: I0321 08:31:46.116803 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 21 08:31:46 crc kubenswrapper[4696]: I0321 08:31:46.299523 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 21 08:31:46 crc kubenswrapper[4696]: I0321 08:31:46.464919 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 21 08:31:46 crc kubenswrapper[4696]: I0321 08:31:46.492970 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 21 08:31:46 crc kubenswrapper[4696]: I0321 08:31:46.507507 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 21 08:31:46 crc kubenswrapper[4696]: I0321 08:31:46.523043 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 21 08:31:46 crc kubenswrapper[4696]: I0321 08:31:46.678943 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 21 08:31:46 crc kubenswrapper[4696]: I0321 08:31:46.688109 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 21 08:31:46 crc kubenswrapper[4696]: I0321 08:31:46.780469 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 21 08:31:46 crc kubenswrapper[4696]: I0321 08:31:46.863465 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 21 08:31:46 crc kubenswrapper[4696]: I0321 08:31:46.906918 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 21 08:31:46 crc kubenswrapper[4696]: I0321 08:31:46.927696 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 21 08:31:47 crc kubenswrapper[4696]: I0321 08:31:47.134844 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 21 08:31:47 crc kubenswrapper[4696]: I0321 08:31:47.201100 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 21 08:31:47 crc kubenswrapper[4696]: I0321 08:31:47.374566 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 21 08:31:47 crc kubenswrapper[4696]: I0321 08:31:47.702400 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 21 08:31:47 crc kubenswrapper[4696]: I0321 08:31:47.717062 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 21 08:31:47 crc kubenswrapper[4696]: I0321 08:31:47.755226 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 21 08:31:47 crc kubenswrapper[4696]: I0321 08:31:47.771138 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 21 08:31:47 crc kubenswrapper[4696]: I0321 08:31:47.906660 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 21 08:31:47 crc kubenswrapper[4696]: I0321 08:31:47.960698 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 21 08:31:48 crc kubenswrapper[4696]: I0321 08:31:48.018984 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 21 08:31:48 crc kubenswrapper[4696]: I0321 08:31:48.034006 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 21 08:31:48 crc kubenswrapper[4696]: I0321 08:31:48.123327 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 21 08:31:48 crc kubenswrapper[4696]: I0321 08:31:48.134517 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 21 08:31:48 crc kubenswrapper[4696]: I0321 08:31:48.134954 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 21 08:31:48 crc kubenswrapper[4696]: I0321 08:31:48.140435 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 21 08:31:48 crc kubenswrapper[4696]: I0321 08:31:48.158329 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 21 08:31:48 crc kubenswrapper[4696]: I0321 08:31:48.246814 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 21 08:31:48 crc kubenswrapper[4696]: I0321 08:31:48.343206 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 21 08:31:48 crc kubenswrapper[4696]: I0321 08:31:48.391956 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 21 08:31:48 crc kubenswrapper[4696]: I0321 08:31:48.410968 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 21 08:31:48 crc kubenswrapper[4696]: I0321 08:31:48.432294 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 21 08:31:48 crc kubenswrapper[4696]: I0321 08:31:48.603920 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 21 08:31:48 crc kubenswrapper[4696]: I0321 08:31:48.675782 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 21 08:31:48 crc kubenswrapper[4696]: I0321 08:31:48.727569 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 21 08:31:48 crc kubenswrapper[4696]: I0321 08:31:48.738967 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 21 08:31:48 crc kubenswrapper[4696]: I0321 08:31:48.788661 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 21 08:31:48 crc kubenswrapper[4696]: I0321 08:31:48.939008 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 21 08:31:49 crc kubenswrapper[4696]: I0321 08:31:49.005148 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 21 08:31:49 crc kubenswrapper[4696]: I0321 08:31:49.026046 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 21 08:31:49 crc kubenswrapper[4696]: I0321 08:31:49.101610 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 21 08:31:49 crc kubenswrapper[4696]: I0321 08:31:49.129236 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 21 08:31:49 crc kubenswrapper[4696]: I0321 08:31:49.172402 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 21 08:31:49 crc kubenswrapper[4696]: I0321 08:31:49.194761 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 21 08:31:49 crc kubenswrapper[4696]: I0321 08:31:49.246505 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 21 08:31:49 crc kubenswrapper[4696]: I0321 08:31:49.292779 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 21 08:31:49 crc kubenswrapper[4696]: I0321 08:31:49.558927 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 21 08:31:49 crc kubenswrapper[4696]: I0321 08:31:49.765316 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 21 08:31:49 crc kubenswrapper[4696]: I0321 08:31:49.851527 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 21 08:31:49 crc kubenswrapper[4696]: I0321 08:31:49.945681 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 21 08:31:49 crc kubenswrapper[4696]: I0321 08:31:49.954947 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.010342 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.022738 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.053695 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.105750 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.107384 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.155639 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.172274 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.172415 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.175480 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.199831 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.200941 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.211608 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.243988 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.255355 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.463188 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.610722 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.702053 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.754093 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.763584 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.890917 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.893546 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.983468 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 21 08:31:50 crc kubenswrapper[4696]: I0321 08:31:50.999012 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 21 08:31:51 crc kubenswrapper[4696]: I0321 08:31:51.060192 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 21 08:31:51 crc kubenswrapper[4696]: I0321 08:31:51.268787 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 21 08:31:51 crc kubenswrapper[4696]: I0321 08:31:51.281267 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 21 08:31:51 crc kubenswrapper[4696]: I0321 08:31:51.467689 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 21 08:31:51 crc kubenswrapper[4696]: I0321 08:31:51.474519 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 21 08:31:51 crc kubenswrapper[4696]: I0321 08:31:51.479266 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 21 08:31:51 crc kubenswrapper[4696]: I0321 08:31:51.519492 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 21 08:31:51 crc kubenswrapper[4696]: I0321 08:31:51.578861 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 21 08:31:51 crc kubenswrapper[4696]: I0321 08:31:51.676411 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 21 08:31:51 crc kubenswrapper[4696]: I0321 08:31:51.761357 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 21 08:31:51 crc kubenswrapper[4696]: I0321 08:31:51.971915 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.110876 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.116171 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.151337 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.246590 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.283231 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.363674 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.386909 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.401987 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.408483 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.418720 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.553615 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.562268 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.575747 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.611755 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.641561 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.655017 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.655919 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.662646 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.678769 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.704751 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.711895 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.723509 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.773213 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.797931 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.868671 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 21 08:31:52 crc kubenswrapper[4696]: I0321 08:31:52.918266 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 21 08:31:53 crc kubenswrapper[4696]: I0321 08:31:53.016385 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 21 08:31:53 crc kubenswrapper[4696]: I0321 08:31:53.020728 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 21 08:31:53 crc kubenswrapper[4696]: I0321 08:31:53.040062 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 21 08:31:53 crc kubenswrapper[4696]: I0321 08:31:53.073756 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 21 08:31:53 crc kubenswrapper[4696]: I0321 08:31:53.151715 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 21 08:31:53 crc kubenswrapper[4696]: I0321 08:31:53.170925 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 21 08:31:53 crc kubenswrapper[4696]: I0321 08:31:53.191038 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 21 08:31:53 crc kubenswrapper[4696]: I0321 08:31:53.230277 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 21 08:31:53 crc kubenswrapper[4696]: I0321 08:31:53.313116 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 21 08:31:53 crc kubenswrapper[4696]: I0321 08:31:53.320931 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 21 08:31:53 crc kubenswrapper[4696]: I0321 08:31:53.619962 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 21 08:31:53 crc kubenswrapper[4696]: I0321 08:31:53.739894 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 21 08:31:53 crc kubenswrapper[4696]: I0321 08:31:53.752176 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 21 08:31:53 crc kubenswrapper[4696]: I0321 08:31:53.753083 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 21 08:31:53 crc kubenswrapper[4696]: I0321 08:31:53.797000 4696 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 21 08:31:53 crc kubenswrapper[4696]: I0321 08:31:53.876576 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 21 08:31:54 crc kubenswrapper[4696]: I0321 08:31:54.030275 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 21 08:31:54 crc kubenswrapper[4696]: I0321 08:31:54.155505 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 21 08:31:54 crc kubenswrapper[4696]: I0321 08:31:54.320353 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 21 08:31:54 crc kubenswrapper[4696]: I0321 08:31:54.346639 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 21 08:31:54 crc kubenswrapper[4696]: I0321 08:31:54.353183 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 21 08:31:54 crc kubenswrapper[4696]: I0321 08:31:54.482976 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 21 08:31:54 crc kubenswrapper[4696]: I0321 08:31:54.485441 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 21 08:31:54 crc kubenswrapper[4696]: I0321 08:31:54.522954 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 21 08:31:54 crc kubenswrapper[4696]: I0321 08:31:54.533315 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 21 08:31:54 crc kubenswrapper[4696]: I0321 08:31:54.718578 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 21 08:31:54 crc kubenswrapper[4696]: I0321 08:31:54.865096 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 21 08:31:54 crc kubenswrapper[4696]: I0321 08:31:54.868570 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 21 08:31:54 crc kubenswrapper[4696]: I0321 08:31:54.875005 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 21 08:31:54 crc kubenswrapper[4696]: I0321 08:31:54.889850 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 21 08:31:54 crc kubenswrapper[4696]: I0321 08:31:54.957542 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 21 08:31:54 crc kubenswrapper[4696]: I0321 08:31:54.981883 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 21 08:31:54 crc kubenswrapper[4696]: I0321 08:31:54.988744 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.054618 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.069020 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.138620 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.142329 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.153363 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.221514 4696 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.271635 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.304431 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.404066 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.438152 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.496176 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.541276 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.550806 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.576414 4696 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.581184 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=43.581163726 podStartE2EDuration="43.581163726s" podCreationTimestamp="2026-03-21 08:31:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:31:31.636948067 +0000 UTC m=+225.757828790" watchObservedRunningTime="2026-03-21 08:31:55.581163726 +0000 UTC m=+249.702044459" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.582031 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.582080 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.587957 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.593853 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.611659 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.613165 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.613483 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=24.613444407 podStartE2EDuration="24.613444407s" podCreationTimestamp="2026-03-21 08:31:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:31:55.604979972 +0000 UTC m=+249.725860715" watchObservedRunningTime="2026-03-21 08:31:55.613444407 +0000 UTC m=+249.734325160" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.678035 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.694096 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.760049 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.765271 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.828468 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.852127 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.904576 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.906523 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.907451 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 21 08:31:55 crc kubenswrapper[4696]: I0321 08:31:55.996058 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 21 08:31:56 crc kubenswrapper[4696]: I0321 08:31:56.013004 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 21 08:31:56 crc kubenswrapper[4696]: I0321 08:31:56.139393 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 21 08:31:56 crc kubenswrapper[4696]: I0321 08:31:56.221871 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 21 08:31:56 crc kubenswrapper[4696]: I0321 08:31:56.252284 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 21 08:31:56 crc kubenswrapper[4696]: I0321 08:31:56.256903 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 21 08:31:56 crc kubenswrapper[4696]: I0321 08:31:56.313285 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 21 08:31:56 crc kubenswrapper[4696]: I0321 08:31:56.375090 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 21 08:31:56 crc kubenswrapper[4696]: I0321 08:31:56.406029 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 21 08:31:56 crc kubenswrapper[4696]: I0321 08:31:56.424310 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 21 08:31:56 crc kubenswrapper[4696]: I0321 08:31:56.475969 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 21 08:31:56 crc kubenswrapper[4696]: I0321 08:31:56.567954 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 21 08:31:56 crc kubenswrapper[4696]: I0321 08:31:56.633008 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 21 08:31:56 crc kubenswrapper[4696]: I0321 08:31:56.693006 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 21 08:31:56 crc kubenswrapper[4696]: I0321 08:31:56.759913 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 21 08:31:56 crc kubenswrapper[4696]: I0321 08:31:56.769453 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 21 08:31:56 crc kubenswrapper[4696]: I0321 08:31:56.970753 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 21 08:31:57 crc kubenswrapper[4696]: I0321 08:31:57.027160 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 21 08:31:57 crc kubenswrapper[4696]: I0321 08:31:57.182702 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 21 08:31:57 crc kubenswrapper[4696]: I0321 08:31:57.232968 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 21 08:31:57 crc kubenswrapper[4696]: I0321 08:31:57.324023 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 21 08:31:57 crc kubenswrapper[4696]: I0321 08:31:57.361097 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 21 08:31:57 crc kubenswrapper[4696]: I0321 08:31:57.385611 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 21 08:31:57 crc kubenswrapper[4696]: I0321 08:31:57.409339 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 21 08:31:57 crc kubenswrapper[4696]: I0321 08:31:57.456350 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 21 08:31:57 crc kubenswrapper[4696]: I0321 08:31:57.654693 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 21 08:31:57 crc kubenswrapper[4696]: I0321 08:31:57.759131 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 21 08:31:57 crc kubenswrapper[4696]: I0321 08:31:57.815568 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 21 08:31:57 crc kubenswrapper[4696]: I0321 08:31:57.882449 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 21 08:31:57 crc kubenswrapper[4696]: I0321 08:31:57.886683 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 21 08:31:57 crc kubenswrapper[4696]: I0321 08:31:57.888442 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 21 08:31:57 crc kubenswrapper[4696]: I0321 08:31:57.890960 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 21 08:31:58 crc kubenswrapper[4696]: I0321 08:31:58.014516 4696 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 21 08:31:58 crc kubenswrapper[4696]: I0321 08:31:58.078420 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 21 08:31:58 crc kubenswrapper[4696]: I0321 08:31:58.134151 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 21 08:31:58 crc kubenswrapper[4696]: I0321 08:31:58.143482 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 21 08:31:58 crc kubenswrapper[4696]: I0321 08:31:58.144980 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 21 08:31:58 crc kubenswrapper[4696]: I0321 08:31:58.227154 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 21 08:31:58 crc kubenswrapper[4696]: I0321 08:31:58.239743 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 21 08:31:58 crc kubenswrapper[4696]: I0321 08:31:58.400678 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 21 08:31:58 crc kubenswrapper[4696]: I0321 08:31:58.489238 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 21 08:31:58 crc kubenswrapper[4696]: I0321 08:31:58.555559 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 21 08:31:58 crc kubenswrapper[4696]: I0321 08:31:58.650890 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 21 08:31:58 crc kubenswrapper[4696]: I0321 08:31:58.755072 4696 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 21 08:31:58 crc kubenswrapper[4696]: I0321 08:31:58.791612 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 21 08:31:58 crc kubenswrapper[4696]: I0321 08:31:58.827551 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 21 08:31:59 crc kubenswrapper[4696]: I0321 08:31:59.140306 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 21 08:31:59 crc kubenswrapper[4696]: I0321 08:31:59.166468 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 21 08:31:59 crc kubenswrapper[4696]: I0321 08:31:59.266212 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 21 08:31:59 crc kubenswrapper[4696]: I0321 08:31:59.341709 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 21 08:31:59 crc kubenswrapper[4696]: I0321 08:31:59.342660 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 21 08:31:59 crc kubenswrapper[4696]: I0321 08:31:59.369750 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 21 08:31:59 crc kubenswrapper[4696]: I0321 08:31:59.496994 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 21 08:31:59 crc kubenswrapper[4696]: I0321 08:31:59.499786 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 21 08:31:59 crc kubenswrapper[4696]: I0321 08:31:59.530694 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 21 08:31:59 crc kubenswrapper[4696]: I0321 08:31:59.871599 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 21 08:32:00 crc kubenswrapper[4696]: I0321 08:32:00.164650 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568032-ljb7d"] Mar 21 08:32:00 crc kubenswrapper[4696]: E0321 08:32:00.164857 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df078e35-cb4d-4be7-beaf-7fd6fc463a0c" containerName="installer" Mar 21 08:32:00 crc kubenswrapper[4696]: I0321 08:32:00.164868 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="df078e35-cb4d-4be7-beaf-7fd6fc463a0c" containerName="installer" Mar 21 08:32:00 crc kubenswrapper[4696]: I0321 08:32:00.164952 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="df078e35-cb4d-4be7-beaf-7fd6fc463a0c" containerName="installer" Mar 21 08:32:00 crc kubenswrapper[4696]: I0321 08:32:00.165269 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568032-ljb7d" Mar 21 08:32:00 crc kubenswrapper[4696]: I0321 08:32:00.169319 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 08:32:00 crc kubenswrapper[4696]: I0321 08:32:00.169319 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 08:32:00 crc kubenswrapper[4696]: I0321 08:32:00.169604 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 08:32:00 crc kubenswrapper[4696]: I0321 08:32:00.252148 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx262\" (UniqueName: \"kubernetes.io/projected/b435a39c-eb1b-4ae9-9cc3-7dc51f632111-kube-api-access-dx262\") pod \"auto-csr-approver-29568032-ljb7d\" (UID: \"b435a39c-eb1b-4ae9-9cc3-7dc51f632111\") " pod="openshift-infra/auto-csr-approver-29568032-ljb7d" Mar 21 08:32:00 crc kubenswrapper[4696]: I0321 08:32:00.341545 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:32:00 crc kubenswrapper[4696]: I0321 08:32:00.342195 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:32:00 crc kubenswrapper[4696]: I0321 08:32:00.349720 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 21 08:32:00 crc kubenswrapper[4696]: I0321 08:32:00.353785 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx262\" (UniqueName: \"kubernetes.io/projected/b435a39c-eb1b-4ae9-9cc3-7dc51f632111-kube-api-access-dx262\") pod \"auto-csr-approver-29568032-ljb7d\" (UID: \"b435a39c-eb1b-4ae9-9cc3-7dc51f632111\") " pod="openshift-infra/auto-csr-approver-29568032-ljb7d" Mar 21 08:32:00 crc kubenswrapper[4696]: I0321 08:32:00.380078 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx262\" (UniqueName: \"kubernetes.io/projected/b435a39c-eb1b-4ae9-9cc3-7dc51f632111-kube-api-access-dx262\") pod \"auto-csr-approver-29568032-ljb7d\" (UID: \"b435a39c-eb1b-4ae9-9cc3-7dc51f632111\") " pod="openshift-infra/auto-csr-approver-29568032-ljb7d" Mar 21 08:32:00 crc kubenswrapper[4696]: I0321 08:32:00.388706 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568032-ljb7d"] Mar 21 08:32:00 crc kubenswrapper[4696]: I0321 08:32:00.479277 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568032-ljb7d" Mar 21 08:32:00 crc kubenswrapper[4696]: W0321 08:32:00.914836 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb435a39c_eb1b_4ae9_9cc3_7dc51f632111.slice/crio-6d2d98a7e2bfb840d1725130cc7a386ef9b60ac3d921ded2f9702457218f1382 WatchSource:0}: Error finding container 6d2d98a7e2bfb840d1725130cc7a386ef9b60ac3d921ded2f9702457218f1382: Status 404 returned error can't find the container with id 6d2d98a7e2bfb840d1725130cc7a386ef9b60ac3d921ded2f9702457218f1382 Mar 21 08:32:00 crc kubenswrapper[4696]: I0321 08:32:00.915445 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568032-ljb7d"] Mar 21 08:32:01 crc kubenswrapper[4696]: I0321 08:32:01.110941 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 21 08:32:01 crc kubenswrapper[4696]: I0321 08:32:01.907884 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568032-ljb7d" event={"ID":"b435a39c-eb1b-4ae9-9cc3-7dc51f632111","Type":"ContainerStarted","Data":"6d2d98a7e2bfb840d1725130cc7a386ef9b60ac3d921ded2f9702457218f1382"} Mar 21 08:32:02 crc kubenswrapper[4696]: I0321 08:32:02.913920 4696 generic.go:334] "Generic (PLEG): container finished" podID="b435a39c-eb1b-4ae9-9cc3-7dc51f632111" containerID="4a956d1c59ecd49c327f9b814d7d2c78baf1cb3ff0693847794badb2a86f9dd7" exitCode=0 Mar 21 08:32:02 crc kubenswrapper[4696]: I0321 08:32:02.913974 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568032-ljb7d" event={"ID":"b435a39c-eb1b-4ae9-9cc3-7dc51f632111","Type":"ContainerDied","Data":"4a956d1c59ecd49c327f9b814d7d2c78baf1cb3ff0693847794badb2a86f9dd7"} Mar 21 08:32:04 crc kubenswrapper[4696]: I0321 08:32:04.238207 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568032-ljb7d" Mar 21 08:32:04 crc kubenswrapper[4696]: I0321 08:32:04.407193 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dx262\" (UniqueName: \"kubernetes.io/projected/b435a39c-eb1b-4ae9-9cc3-7dc51f632111-kube-api-access-dx262\") pod \"b435a39c-eb1b-4ae9-9cc3-7dc51f632111\" (UID: \"b435a39c-eb1b-4ae9-9cc3-7dc51f632111\") " Mar 21 08:32:04 crc kubenswrapper[4696]: I0321 08:32:04.414570 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b435a39c-eb1b-4ae9-9cc3-7dc51f632111-kube-api-access-dx262" (OuterVolumeSpecName: "kube-api-access-dx262") pod "b435a39c-eb1b-4ae9-9cc3-7dc51f632111" (UID: "b435a39c-eb1b-4ae9-9cc3-7dc51f632111"). InnerVolumeSpecName "kube-api-access-dx262". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:32:04 crc kubenswrapper[4696]: I0321 08:32:04.508667 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dx262\" (UniqueName: \"kubernetes.io/projected/b435a39c-eb1b-4ae9-9cc3-7dc51f632111-kube-api-access-dx262\") on node \"crc\" DevicePath \"\"" Mar 21 08:32:04 crc kubenswrapper[4696]: I0321 08:32:04.924917 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568032-ljb7d" event={"ID":"b435a39c-eb1b-4ae9-9cc3-7dc51f632111","Type":"ContainerDied","Data":"6d2d98a7e2bfb840d1725130cc7a386ef9b60ac3d921ded2f9702457218f1382"} Mar 21 08:32:04 crc kubenswrapper[4696]: I0321 08:32:04.924949 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d2d98a7e2bfb840d1725130cc7a386ef9b60ac3d921ded2f9702457218f1382" Mar 21 08:32:04 crc kubenswrapper[4696]: I0321 08:32:04.925025 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568032-ljb7d" Mar 21 08:32:05 crc kubenswrapper[4696]: I0321 08:32:05.331964 4696 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 21 08:32:05 crc kubenswrapper[4696]: I0321 08:32:05.332168 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://d5becd6a47f3d578519a657f70c3a6a05c41c399caeb6841448da9b2df4fac5c" gracePeriod=5 Mar 21 08:32:10 crc kubenswrapper[4696]: I0321 08:32:10.925020 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 21 08:32:10 crc kubenswrapper[4696]: I0321 08:32:10.926182 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 08:32:10 crc kubenswrapper[4696]: I0321 08:32:10.954438 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 21 08:32:10 crc kubenswrapper[4696]: I0321 08:32:10.954684 4696 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="d5becd6a47f3d578519a657f70c3a6a05c41c399caeb6841448da9b2df4fac5c" exitCode=137 Mar 21 08:32:10 crc kubenswrapper[4696]: I0321 08:32:10.954840 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 08:32:10 crc kubenswrapper[4696]: I0321 08:32:10.954770 4696 scope.go:117] "RemoveContainer" containerID="d5becd6a47f3d578519a657f70c3a6a05c41c399caeb6841448da9b2df4fac5c" Mar 21 08:32:10 crc kubenswrapper[4696]: I0321 08:32:10.974374 4696 scope.go:117] "RemoveContainer" containerID="d5becd6a47f3d578519a657f70c3a6a05c41c399caeb6841448da9b2df4fac5c" Mar 21 08:32:10 crc kubenswrapper[4696]: E0321 08:32:10.974867 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5becd6a47f3d578519a657f70c3a6a05c41c399caeb6841448da9b2df4fac5c\": container with ID starting with d5becd6a47f3d578519a657f70c3a6a05c41c399caeb6841448da9b2df4fac5c not found: ID does not exist" containerID="d5becd6a47f3d578519a657f70c3a6a05c41c399caeb6841448da9b2df4fac5c" Mar 21 08:32:10 crc kubenswrapper[4696]: I0321 08:32:10.975002 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5becd6a47f3d578519a657f70c3a6a05c41c399caeb6841448da9b2df4fac5c"} err="failed to get container status \"d5becd6a47f3d578519a657f70c3a6a05c41c399caeb6841448da9b2df4fac5c\": rpc error: code = NotFound desc = could not find container \"d5becd6a47f3d578519a657f70c3a6a05c41c399caeb6841448da9b2df4fac5c\": container with ID starting with d5becd6a47f3d578519a657f70c3a6a05c41c399caeb6841448da9b2df4fac5c not found: ID does not exist" Mar 21 08:32:11 crc kubenswrapper[4696]: I0321 08:32:11.087425 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 21 08:32:11 crc kubenswrapper[4696]: I0321 08:32:11.087492 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 21 08:32:11 crc kubenswrapper[4696]: I0321 08:32:11.087545 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 21 08:32:11 crc kubenswrapper[4696]: I0321 08:32:11.087624 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 21 08:32:11 crc kubenswrapper[4696]: I0321 08:32:11.087646 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 21 08:32:11 crc kubenswrapper[4696]: I0321 08:32:11.087686 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:32:11 crc kubenswrapper[4696]: I0321 08:32:11.087747 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:32:11 crc kubenswrapper[4696]: I0321 08:32:11.087791 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:32:11 crc kubenswrapper[4696]: I0321 08:32:11.087850 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:32:11 crc kubenswrapper[4696]: I0321 08:32:11.088058 4696 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 21 08:32:11 crc kubenswrapper[4696]: I0321 08:32:11.088076 4696 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 21 08:32:11 crc kubenswrapper[4696]: I0321 08:32:11.088087 4696 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 21 08:32:11 crc kubenswrapper[4696]: I0321 08:32:11.088098 4696 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 21 08:32:11 crc kubenswrapper[4696]: I0321 08:32:11.095689 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:32:11 crc kubenswrapper[4696]: I0321 08:32:11.189157 4696 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 21 08:32:12 crc kubenswrapper[4696]: I0321 08:32:12.548186 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 21 08:32:12 crc kubenswrapper[4696]: I0321 08:32:12.548467 4696 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Mar 21 08:32:12 crc kubenswrapper[4696]: I0321 08:32:12.564026 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 21 08:32:12 crc kubenswrapper[4696]: I0321 08:32:12.564087 4696 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="8b0a6416-af56-4378-aa9f-144f3973db77" Mar 21 08:32:12 crc kubenswrapper[4696]: I0321 08:32:12.571908 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 21 08:32:12 crc kubenswrapper[4696]: I0321 08:32:12.572095 4696 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="8b0a6416-af56-4378-aa9f-144f3973db77" Mar 21 08:32:30 crc kubenswrapper[4696]: I0321 08:32:30.341353 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:32:30 crc kubenswrapper[4696]: I0321 08:32:30.341932 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:32:30 crc kubenswrapper[4696]: I0321 08:32:30.341985 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:32:30 crc kubenswrapper[4696]: I0321 08:32:30.342638 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9606dfcaf248d5014358578adf21a000d5fb5d698e3980a6ae0ec673dfb6ef85"} pod="openshift-machine-config-operator/machine-config-daemon-z7srw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 08:32:30 crc kubenswrapper[4696]: I0321 08:32:30.342705 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" containerID="cri-o://9606dfcaf248d5014358578adf21a000d5fb5d698e3980a6ae0ec673dfb6ef85" gracePeriod=600 Mar 21 08:32:31 crc kubenswrapper[4696]: I0321 08:32:31.063678 4696 generic.go:334] "Generic (PLEG): container finished" podID="daaf227a-2305-495e-8495-a280abcd8e10" containerID="9606dfcaf248d5014358578adf21a000d5fb5d698e3980a6ae0ec673dfb6ef85" exitCode=0 Mar 21 08:32:31 crc kubenswrapper[4696]: I0321 08:32:31.063762 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerDied","Data":"9606dfcaf248d5014358578adf21a000d5fb5d698e3980a6ae0ec673dfb6ef85"} Mar 21 08:32:31 crc kubenswrapper[4696]: I0321 08:32:31.064559 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerStarted","Data":"89c4ac42755bc23804bb46eebf22b6cb1af3d359889a73dd951e91867e71e087"} Mar 21 08:33:12 crc kubenswrapper[4696]: I0321 08:33:12.867645 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fkf7p"] Mar 21 08:33:12 crc kubenswrapper[4696]: E0321 08:33:12.868199 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 21 08:33:12 crc kubenswrapper[4696]: I0321 08:33:12.868212 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 21 08:33:12 crc kubenswrapper[4696]: E0321 08:33:12.868234 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b435a39c-eb1b-4ae9-9cc3-7dc51f632111" containerName="oc" Mar 21 08:33:12 crc kubenswrapper[4696]: I0321 08:33:12.868242 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b435a39c-eb1b-4ae9-9cc3-7dc51f632111" containerName="oc" Mar 21 08:33:12 crc kubenswrapper[4696]: I0321 08:33:12.868379 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 21 08:33:12 crc kubenswrapper[4696]: I0321 08:33:12.868390 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="b435a39c-eb1b-4ae9-9cc3-7dc51f632111" containerName="oc" Mar 21 08:33:12 crc kubenswrapper[4696]: I0321 08:33:12.868750 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:12 crc kubenswrapper[4696]: I0321 08:33:12.890594 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fkf7p"] Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.013216 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4b1083de-6a60-4546-baaf-ba42e02ebfd0-registry-certificates\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.013261 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4b1083de-6a60-4546-baaf-ba42e02ebfd0-trusted-ca\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.013280 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4b1083de-6a60-4546-baaf-ba42e02ebfd0-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.013304 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4b1083de-6a60-4546-baaf-ba42e02ebfd0-registry-tls\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.013339 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.013356 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4b1083de-6a60-4546-baaf-ba42e02ebfd0-bound-sa-token\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.013375 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5tjp\" (UniqueName: \"kubernetes.io/projected/4b1083de-6a60-4546-baaf-ba42e02ebfd0-kube-api-access-b5tjp\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.013404 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4b1083de-6a60-4546-baaf-ba42e02ebfd0-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.032328 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.114093 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4b1083de-6a60-4546-baaf-ba42e02ebfd0-bound-sa-token\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.114127 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5tjp\" (UniqueName: \"kubernetes.io/projected/4b1083de-6a60-4546-baaf-ba42e02ebfd0-kube-api-access-b5tjp\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.114158 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4b1083de-6a60-4546-baaf-ba42e02ebfd0-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.114191 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4b1083de-6a60-4546-baaf-ba42e02ebfd0-registry-certificates\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.114217 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4b1083de-6a60-4546-baaf-ba42e02ebfd0-trusted-ca\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.114235 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4b1083de-6a60-4546-baaf-ba42e02ebfd0-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.114260 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4b1083de-6a60-4546-baaf-ba42e02ebfd0-registry-tls\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.115374 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4b1083de-6a60-4546-baaf-ba42e02ebfd0-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.115731 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4b1083de-6a60-4546-baaf-ba42e02ebfd0-trusted-ca\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.115769 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4b1083de-6a60-4546-baaf-ba42e02ebfd0-registry-certificates\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.119681 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4b1083de-6a60-4546-baaf-ba42e02ebfd0-registry-tls\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.121188 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4b1083de-6a60-4546-baaf-ba42e02ebfd0-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.128841 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4b1083de-6a60-4546-baaf-ba42e02ebfd0-bound-sa-token\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.132993 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5tjp\" (UniqueName: \"kubernetes.io/projected/4b1083de-6a60-4546-baaf-ba42e02ebfd0-kube-api-access-b5tjp\") pod \"image-registry-66df7c8f76-fkf7p\" (UID: \"4b1083de-6a60-4546-baaf-ba42e02ebfd0\") " pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.182369 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:13 crc kubenswrapper[4696]: I0321 08:33:13.371296 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fkf7p"] Mar 21 08:33:14 crc kubenswrapper[4696]: I0321 08:33:14.304027 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" event={"ID":"4b1083de-6a60-4546-baaf-ba42e02ebfd0","Type":"ContainerStarted","Data":"c907e304caabf6b615f5d57e99bf40e8e74c07909a836d9bdd26c4f516c90740"} Mar 21 08:33:14 crc kubenswrapper[4696]: I0321 08:33:14.304480 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:14 crc kubenswrapper[4696]: I0321 08:33:14.304512 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" event={"ID":"4b1083de-6a60-4546-baaf-ba42e02ebfd0","Type":"ContainerStarted","Data":"a7e61af595045642d9523d8e04958a886c4ffe36d3b25dd01b6d1639af36f978"} Mar 21 08:33:14 crc kubenswrapper[4696]: I0321 08:33:14.331969 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" podStartSLOduration=2.331942185 podStartE2EDuration="2.331942185s" podCreationTimestamp="2026-03-21 08:33:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:33:14.323307162 +0000 UTC m=+328.444187905" watchObservedRunningTime="2026-03-21 08:33:14.331942185 +0000 UTC m=+328.452822938" Mar 21 08:33:25 crc kubenswrapper[4696]: I0321 08:33:25.877046 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qsg67"] Mar 21 08:33:25 crc kubenswrapper[4696]: I0321 08:33:25.887914 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v9rlg"] Mar 21 08:33:25 crc kubenswrapper[4696]: I0321 08:33:25.888241 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v9rlg" podUID="85a5000b-3b68-4808-bcfc-93bcbde10dde" containerName="registry-server" containerID="cri-o://883148dde23dfd1d81101717644f6ddfa087aafb6d6f2a0659071c39c883dfbe" gracePeriod=30 Mar 21 08:33:25 crc kubenswrapper[4696]: I0321 08:33:25.900779 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5vrbx"] Mar 21 08:33:25 crc kubenswrapper[4696]: I0321 08:33:25.901261 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" podUID="973fa2b5-c6ce-498f-9b2b-311e8dd34057" containerName="marketplace-operator" containerID="cri-o://6ed0dd17f07cf887f320e3f6560876112a3577e3274a4b0d84c8e65baede4294" gracePeriod=30 Mar 21 08:33:25 crc kubenswrapper[4696]: I0321 08:33:25.905489 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jnw8m"] Mar 21 08:33:25 crc kubenswrapper[4696]: I0321 08:33:25.905762 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jnw8m" podUID="d70a46f0-1e5d-4edb-8745-9b8a41b66241" containerName="registry-server" containerID="cri-o://ad76337a74c84f6126a24fc0768cd55ff27a998a9d26ef3c78b7b12073b148c3" gracePeriod=30 Mar 21 08:33:25 crc kubenswrapper[4696]: I0321 08:33:25.911280 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cfjs8"] Mar 21 08:33:25 crc kubenswrapper[4696]: I0321 08:33:25.911567 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cfjs8" podUID="0030cc8a-f473-4d0a-ace9-728b66e62634" containerName="registry-server" containerID="cri-o://3ce447a4777038b8616bc6dd26064b084fa129af52ead66b5221e39c940bdc0c" gracePeriod=30 Mar 21 08:33:25 crc kubenswrapper[4696]: I0321 08:33:25.915739 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w6xbt"] Mar 21 08:33:25 crc kubenswrapper[4696]: I0321 08:33:25.917119 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-w6xbt" Mar 21 08:33:25 crc kubenswrapper[4696]: I0321 08:33:25.946046 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w6xbt"] Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.008303 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d45283e8-56c7-4a2e-8a7e-88e459bb1c04-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-w6xbt\" (UID: \"d45283e8-56c7-4a2e-8a7e-88e459bb1c04\") " pod="openshift-marketplace/marketplace-operator-79b997595-w6xbt" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.008610 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9srs\" (UniqueName: \"kubernetes.io/projected/d45283e8-56c7-4a2e-8a7e-88e459bb1c04-kube-api-access-x9srs\") pod \"marketplace-operator-79b997595-w6xbt\" (UID: \"d45283e8-56c7-4a2e-8a7e-88e459bb1c04\") " pod="openshift-marketplace/marketplace-operator-79b997595-w6xbt" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.008725 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d45283e8-56c7-4a2e-8a7e-88e459bb1c04-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-w6xbt\" (UID: \"d45283e8-56c7-4a2e-8a7e-88e459bb1c04\") " pod="openshift-marketplace/marketplace-operator-79b997595-w6xbt" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.110320 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9srs\" (UniqueName: \"kubernetes.io/projected/d45283e8-56c7-4a2e-8a7e-88e459bb1c04-kube-api-access-x9srs\") pod \"marketplace-operator-79b997595-w6xbt\" (UID: \"d45283e8-56c7-4a2e-8a7e-88e459bb1c04\") " pod="openshift-marketplace/marketplace-operator-79b997595-w6xbt" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.110404 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d45283e8-56c7-4a2e-8a7e-88e459bb1c04-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-w6xbt\" (UID: \"d45283e8-56c7-4a2e-8a7e-88e459bb1c04\") " pod="openshift-marketplace/marketplace-operator-79b997595-w6xbt" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.110459 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d45283e8-56c7-4a2e-8a7e-88e459bb1c04-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-w6xbt\" (UID: \"d45283e8-56c7-4a2e-8a7e-88e459bb1c04\") " pod="openshift-marketplace/marketplace-operator-79b997595-w6xbt" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.111706 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d45283e8-56c7-4a2e-8a7e-88e459bb1c04-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-w6xbt\" (UID: \"d45283e8-56c7-4a2e-8a7e-88e459bb1c04\") " pod="openshift-marketplace/marketplace-operator-79b997595-w6xbt" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.116042 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d45283e8-56c7-4a2e-8a7e-88e459bb1c04-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-w6xbt\" (UID: \"d45283e8-56c7-4a2e-8a7e-88e459bb1c04\") " pod="openshift-marketplace/marketplace-operator-79b997595-w6xbt" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.131906 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9srs\" (UniqueName: \"kubernetes.io/projected/d45283e8-56c7-4a2e-8a7e-88e459bb1c04-kube-api-access-x9srs\") pod \"marketplace-operator-79b997595-w6xbt\" (UID: \"d45283e8-56c7-4a2e-8a7e-88e459bb1c04\") " pod="openshift-marketplace/marketplace-operator-79b997595-w6xbt" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.235014 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-w6xbt" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.248221 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cfjs8" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.317353 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n825s\" (UniqueName: \"kubernetes.io/projected/0030cc8a-f473-4d0a-ace9-728b66e62634-kube-api-access-n825s\") pod \"0030cc8a-f473-4d0a-ace9-728b66e62634\" (UID: \"0030cc8a-f473-4d0a-ace9-728b66e62634\") " Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.317449 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0030cc8a-f473-4d0a-ace9-728b66e62634-utilities\") pod \"0030cc8a-f473-4d0a-ace9-728b66e62634\" (UID: \"0030cc8a-f473-4d0a-ace9-728b66e62634\") " Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.317540 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0030cc8a-f473-4d0a-ace9-728b66e62634-catalog-content\") pod \"0030cc8a-f473-4d0a-ace9-728b66e62634\" (UID: \"0030cc8a-f473-4d0a-ace9-728b66e62634\") " Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.318571 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0030cc8a-f473-4d0a-ace9-728b66e62634-utilities" (OuterVolumeSpecName: "utilities") pod "0030cc8a-f473-4d0a-ace9-728b66e62634" (UID: "0030cc8a-f473-4d0a-ace9-728b66e62634"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.318893 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0030cc8a-f473-4d0a-ace9-728b66e62634-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.324134 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0030cc8a-f473-4d0a-ace9-728b66e62634-kube-api-access-n825s" (OuterVolumeSpecName: "kube-api-access-n825s") pod "0030cc8a-f473-4d0a-ace9-728b66e62634" (UID: "0030cc8a-f473-4d0a-ace9-728b66e62634"). InnerVolumeSpecName "kube-api-access-n825s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.351166 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v9rlg" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.353657 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.392707 4696 generic.go:334] "Generic (PLEG): container finished" podID="973fa2b5-c6ce-498f-9b2b-311e8dd34057" containerID="6ed0dd17f07cf887f320e3f6560876112a3577e3274a4b0d84c8e65baede4294" exitCode=0 Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.392866 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" event={"ID":"973fa2b5-c6ce-498f-9b2b-311e8dd34057","Type":"ContainerDied","Data":"6ed0dd17f07cf887f320e3f6560876112a3577e3274a4b0d84c8e65baede4294"} Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.392895 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" event={"ID":"973fa2b5-c6ce-498f-9b2b-311e8dd34057","Type":"ContainerDied","Data":"4f55d748904168540bea4ad08042f6ff645af26914eb9737ac872aaa9d1627ba"} Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.392911 4696 scope.go:117] "RemoveContainer" containerID="6ed0dd17f07cf887f320e3f6560876112a3577e3274a4b0d84c8e65baede4294" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.393025 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5vrbx" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.406159 4696 generic.go:334] "Generic (PLEG): container finished" podID="0030cc8a-f473-4d0a-ace9-728b66e62634" containerID="3ce447a4777038b8616bc6dd26064b084fa129af52ead66b5221e39c940bdc0c" exitCode=0 Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.406218 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cfjs8" event={"ID":"0030cc8a-f473-4d0a-ace9-728b66e62634","Type":"ContainerDied","Data":"3ce447a4777038b8616bc6dd26064b084fa129af52ead66b5221e39c940bdc0c"} Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.406246 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cfjs8" event={"ID":"0030cc8a-f473-4d0a-ace9-728b66e62634","Type":"ContainerDied","Data":"08ec3a1b8717c1c8352f9398afd7a2b7497ef06f68318a7ae0c59170c0aeb3b9"} Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.406300 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cfjs8" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.417734 4696 generic.go:334] "Generic (PLEG): container finished" podID="85a5000b-3b68-4808-bcfc-93bcbde10dde" containerID="883148dde23dfd1d81101717644f6ddfa087aafb6d6f2a0659071c39c883dfbe" exitCode=0 Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.417780 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v9rlg" event={"ID":"85a5000b-3b68-4808-bcfc-93bcbde10dde","Type":"ContainerDied","Data":"883148dde23dfd1d81101717644f6ddfa087aafb6d6f2a0659071c39c883dfbe"} Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.417947 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v9rlg" event={"ID":"85a5000b-3b68-4808-bcfc-93bcbde10dde","Type":"ContainerDied","Data":"404a5067b9555350774dc249e338ff2987f6c02171ba5c83b7c05b916470b110"} Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.417797 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v9rlg" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.421803 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85a5000b-3b68-4808-bcfc-93bcbde10dde-utilities\") pod \"85a5000b-3b68-4808-bcfc-93bcbde10dde\" (UID: \"85a5000b-3b68-4808-bcfc-93bcbde10dde\") " Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.421881 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdk58\" (UniqueName: \"kubernetes.io/projected/973fa2b5-c6ce-498f-9b2b-311e8dd34057-kube-api-access-sdk58\") pod \"973fa2b5-c6ce-498f-9b2b-311e8dd34057\" (UID: \"973fa2b5-c6ce-498f-9b2b-311e8dd34057\") " Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.421959 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/973fa2b5-c6ce-498f-9b2b-311e8dd34057-marketplace-operator-metrics\") pod \"973fa2b5-c6ce-498f-9b2b-311e8dd34057\" (UID: \"973fa2b5-c6ce-498f-9b2b-311e8dd34057\") " Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.421999 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zxpn\" (UniqueName: \"kubernetes.io/projected/85a5000b-3b68-4808-bcfc-93bcbde10dde-kube-api-access-4zxpn\") pod \"85a5000b-3b68-4808-bcfc-93bcbde10dde\" (UID: \"85a5000b-3b68-4808-bcfc-93bcbde10dde\") " Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.422033 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85a5000b-3b68-4808-bcfc-93bcbde10dde-catalog-content\") pod \"85a5000b-3b68-4808-bcfc-93bcbde10dde\" (UID: \"85a5000b-3b68-4808-bcfc-93bcbde10dde\") " Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.422084 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/973fa2b5-c6ce-498f-9b2b-311e8dd34057-marketplace-trusted-ca\") pod \"973fa2b5-c6ce-498f-9b2b-311e8dd34057\" (UID: \"973fa2b5-c6ce-498f-9b2b-311e8dd34057\") " Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.422788 4696 generic.go:334] "Generic (PLEG): container finished" podID="d70a46f0-1e5d-4edb-8745-9b8a41b66241" containerID="ad76337a74c84f6126a24fc0768cd55ff27a998a9d26ef3c78b7b12073b148c3" exitCode=0 Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.422971 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85a5000b-3b68-4808-bcfc-93bcbde10dde-utilities" (OuterVolumeSpecName: "utilities") pod "85a5000b-3b68-4808-bcfc-93bcbde10dde" (UID: "85a5000b-3b68-4808-bcfc-93bcbde10dde"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.423000 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qsg67" podUID="6aabaca2-7def-4550-bae1-9337b7c1002b" containerName="registry-server" containerID="cri-o://bdc0c304327cc59729598ff33c42a54e45d127e5d42d5558ef2977e434010ea4" gracePeriod=30 Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.423404 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jnw8m" event={"ID":"d70a46f0-1e5d-4edb-8745-9b8a41b66241","Type":"ContainerDied","Data":"ad76337a74c84f6126a24fc0768cd55ff27a998a9d26ef3c78b7b12073b148c3"} Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.423997 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/973fa2b5-c6ce-498f-9b2b-311e8dd34057-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "973fa2b5-c6ce-498f-9b2b-311e8dd34057" (UID: "973fa2b5-c6ce-498f-9b2b-311e8dd34057"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.424615 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n825s\" (UniqueName: \"kubernetes.io/projected/0030cc8a-f473-4d0a-ace9-728b66e62634-kube-api-access-n825s\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.426924 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85a5000b-3b68-4808-bcfc-93bcbde10dde-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.455508 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/973fa2b5-c6ce-498f-9b2b-311e8dd34057-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "973fa2b5-c6ce-498f-9b2b-311e8dd34057" (UID: "973fa2b5-c6ce-498f-9b2b-311e8dd34057"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.456995 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/973fa2b5-c6ce-498f-9b2b-311e8dd34057-kube-api-access-sdk58" (OuterVolumeSpecName: "kube-api-access-sdk58") pod "973fa2b5-c6ce-498f-9b2b-311e8dd34057" (UID: "973fa2b5-c6ce-498f-9b2b-311e8dd34057"). InnerVolumeSpecName "kube-api-access-sdk58". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.457153 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85a5000b-3b68-4808-bcfc-93bcbde10dde-kube-api-access-4zxpn" (OuterVolumeSpecName: "kube-api-access-4zxpn") pod "85a5000b-3b68-4808-bcfc-93bcbde10dde" (UID: "85a5000b-3b68-4808-bcfc-93bcbde10dde"). InnerVolumeSpecName "kube-api-access-4zxpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.496830 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85a5000b-3b68-4808-bcfc-93bcbde10dde-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "85a5000b-3b68-4808-bcfc-93bcbde10dde" (UID: "85a5000b-3b68-4808-bcfc-93bcbde10dde"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.503707 4696 scope.go:117] "RemoveContainer" containerID="6ed0dd17f07cf887f320e3f6560876112a3577e3274a4b0d84c8e65baede4294" Mar 21 08:33:26 crc kubenswrapper[4696]: E0321 08:33:26.505879 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ed0dd17f07cf887f320e3f6560876112a3577e3274a4b0d84c8e65baede4294\": container with ID starting with 6ed0dd17f07cf887f320e3f6560876112a3577e3274a4b0d84c8e65baede4294 not found: ID does not exist" containerID="6ed0dd17f07cf887f320e3f6560876112a3577e3274a4b0d84c8e65baede4294" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.505908 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ed0dd17f07cf887f320e3f6560876112a3577e3274a4b0d84c8e65baede4294"} err="failed to get container status \"6ed0dd17f07cf887f320e3f6560876112a3577e3274a4b0d84c8e65baede4294\": rpc error: code = NotFound desc = could not find container \"6ed0dd17f07cf887f320e3f6560876112a3577e3274a4b0d84c8e65baede4294\": container with ID starting with 6ed0dd17f07cf887f320e3f6560876112a3577e3274a4b0d84c8e65baede4294 not found: ID does not exist" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.505933 4696 scope.go:117] "RemoveContainer" containerID="3ce447a4777038b8616bc6dd26064b084fa129af52ead66b5221e39c940bdc0c" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.508805 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jnw8m" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.527689 4696 scope.go:117] "RemoveContainer" containerID="a07537b48c01840383ea0d043cf82e3aaf698087df27576814f170898198a1ad" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.538471 4696 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/973fa2b5-c6ce-498f-9b2b-311e8dd34057-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.538520 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdk58\" (UniqueName: \"kubernetes.io/projected/973fa2b5-c6ce-498f-9b2b-311e8dd34057-kube-api-access-sdk58\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.538531 4696 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/973fa2b5-c6ce-498f-9b2b-311e8dd34057-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.538541 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zxpn\" (UniqueName: \"kubernetes.io/projected/85a5000b-3b68-4808-bcfc-93bcbde10dde-kube-api-access-4zxpn\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.538552 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85a5000b-3b68-4808-bcfc-93bcbde10dde-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.564961 4696 scope.go:117] "RemoveContainer" containerID="c7319785d59398994daa2833f5887a3596b380c96b61a56700381ac81bcf2730" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.571052 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0030cc8a-f473-4d0a-ace9-728b66e62634-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0030cc8a-f473-4d0a-ace9-728b66e62634" (UID: "0030cc8a-f473-4d0a-ace9-728b66e62634"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.587098 4696 scope.go:117] "RemoveContainer" containerID="3ce447a4777038b8616bc6dd26064b084fa129af52ead66b5221e39c940bdc0c" Mar 21 08:33:26 crc kubenswrapper[4696]: E0321 08:33:26.587490 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ce447a4777038b8616bc6dd26064b084fa129af52ead66b5221e39c940bdc0c\": container with ID starting with 3ce447a4777038b8616bc6dd26064b084fa129af52ead66b5221e39c940bdc0c not found: ID does not exist" containerID="3ce447a4777038b8616bc6dd26064b084fa129af52ead66b5221e39c940bdc0c" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.587517 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ce447a4777038b8616bc6dd26064b084fa129af52ead66b5221e39c940bdc0c"} err="failed to get container status \"3ce447a4777038b8616bc6dd26064b084fa129af52ead66b5221e39c940bdc0c\": rpc error: code = NotFound desc = could not find container \"3ce447a4777038b8616bc6dd26064b084fa129af52ead66b5221e39c940bdc0c\": container with ID starting with 3ce447a4777038b8616bc6dd26064b084fa129af52ead66b5221e39c940bdc0c not found: ID does not exist" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.587547 4696 scope.go:117] "RemoveContainer" containerID="a07537b48c01840383ea0d043cf82e3aaf698087df27576814f170898198a1ad" Mar 21 08:33:26 crc kubenswrapper[4696]: E0321 08:33:26.589067 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a07537b48c01840383ea0d043cf82e3aaf698087df27576814f170898198a1ad\": container with ID starting with a07537b48c01840383ea0d043cf82e3aaf698087df27576814f170898198a1ad not found: ID does not exist" containerID="a07537b48c01840383ea0d043cf82e3aaf698087df27576814f170898198a1ad" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.589130 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a07537b48c01840383ea0d043cf82e3aaf698087df27576814f170898198a1ad"} err="failed to get container status \"a07537b48c01840383ea0d043cf82e3aaf698087df27576814f170898198a1ad\": rpc error: code = NotFound desc = could not find container \"a07537b48c01840383ea0d043cf82e3aaf698087df27576814f170898198a1ad\": container with ID starting with a07537b48c01840383ea0d043cf82e3aaf698087df27576814f170898198a1ad not found: ID does not exist" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.589165 4696 scope.go:117] "RemoveContainer" containerID="c7319785d59398994daa2833f5887a3596b380c96b61a56700381ac81bcf2730" Mar 21 08:33:26 crc kubenswrapper[4696]: E0321 08:33:26.589441 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7319785d59398994daa2833f5887a3596b380c96b61a56700381ac81bcf2730\": container with ID starting with c7319785d59398994daa2833f5887a3596b380c96b61a56700381ac81bcf2730 not found: ID does not exist" containerID="c7319785d59398994daa2833f5887a3596b380c96b61a56700381ac81bcf2730" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.589473 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7319785d59398994daa2833f5887a3596b380c96b61a56700381ac81bcf2730"} err="failed to get container status \"c7319785d59398994daa2833f5887a3596b380c96b61a56700381ac81bcf2730\": rpc error: code = NotFound desc = could not find container \"c7319785d59398994daa2833f5887a3596b380c96b61a56700381ac81bcf2730\": container with ID starting with c7319785d59398994daa2833f5887a3596b380c96b61a56700381ac81bcf2730 not found: ID does not exist" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.589491 4696 scope.go:117] "RemoveContainer" containerID="883148dde23dfd1d81101717644f6ddfa087aafb6d6f2a0659071c39c883dfbe" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.602219 4696 scope.go:117] "RemoveContainer" containerID="d1d423895bc1e7afe3dda0ff1a30e13f1deafd6c471f6b26594f023d3828ec23" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.613285 4696 scope.go:117] "RemoveContainer" containerID="473840c7910cf1720058db29c86564e69cb039cf4db0ba23a177368175ed1f58" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.640472 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d70a46f0-1e5d-4edb-8745-9b8a41b66241-utilities\") pod \"d70a46f0-1e5d-4edb-8745-9b8a41b66241\" (UID: \"d70a46f0-1e5d-4edb-8745-9b8a41b66241\") " Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.640535 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d70a46f0-1e5d-4edb-8745-9b8a41b66241-catalog-content\") pod \"d70a46f0-1e5d-4edb-8745-9b8a41b66241\" (UID: \"d70a46f0-1e5d-4edb-8745-9b8a41b66241\") " Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.640573 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjjfw\" (UniqueName: \"kubernetes.io/projected/d70a46f0-1e5d-4edb-8745-9b8a41b66241-kube-api-access-fjjfw\") pod \"d70a46f0-1e5d-4edb-8745-9b8a41b66241\" (UID: \"d70a46f0-1e5d-4edb-8745-9b8a41b66241\") " Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.640889 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0030cc8a-f473-4d0a-ace9-728b66e62634-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.641496 4696 scope.go:117] "RemoveContainer" containerID="883148dde23dfd1d81101717644f6ddfa087aafb6d6f2a0659071c39c883dfbe" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.641548 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d70a46f0-1e5d-4edb-8745-9b8a41b66241-utilities" (OuterVolumeSpecName: "utilities") pod "d70a46f0-1e5d-4edb-8745-9b8a41b66241" (UID: "d70a46f0-1e5d-4edb-8745-9b8a41b66241"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:33:26 crc kubenswrapper[4696]: E0321 08:33:26.642126 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"883148dde23dfd1d81101717644f6ddfa087aafb6d6f2a0659071c39c883dfbe\": container with ID starting with 883148dde23dfd1d81101717644f6ddfa087aafb6d6f2a0659071c39c883dfbe not found: ID does not exist" containerID="883148dde23dfd1d81101717644f6ddfa087aafb6d6f2a0659071c39c883dfbe" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.642161 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"883148dde23dfd1d81101717644f6ddfa087aafb6d6f2a0659071c39c883dfbe"} err="failed to get container status \"883148dde23dfd1d81101717644f6ddfa087aafb6d6f2a0659071c39c883dfbe\": rpc error: code = NotFound desc = could not find container \"883148dde23dfd1d81101717644f6ddfa087aafb6d6f2a0659071c39c883dfbe\": container with ID starting with 883148dde23dfd1d81101717644f6ddfa087aafb6d6f2a0659071c39c883dfbe not found: ID does not exist" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.642186 4696 scope.go:117] "RemoveContainer" containerID="d1d423895bc1e7afe3dda0ff1a30e13f1deafd6c471f6b26594f023d3828ec23" Mar 21 08:33:26 crc kubenswrapper[4696]: E0321 08:33:26.642708 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1d423895bc1e7afe3dda0ff1a30e13f1deafd6c471f6b26594f023d3828ec23\": container with ID starting with d1d423895bc1e7afe3dda0ff1a30e13f1deafd6c471f6b26594f023d3828ec23 not found: ID does not exist" containerID="d1d423895bc1e7afe3dda0ff1a30e13f1deafd6c471f6b26594f023d3828ec23" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.642733 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1d423895bc1e7afe3dda0ff1a30e13f1deafd6c471f6b26594f023d3828ec23"} err="failed to get container status \"d1d423895bc1e7afe3dda0ff1a30e13f1deafd6c471f6b26594f023d3828ec23\": rpc error: code = NotFound desc = could not find container \"d1d423895bc1e7afe3dda0ff1a30e13f1deafd6c471f6b26594f023d3828ec23\": container with ID starting with d1d423895bc1e7afe3dda0ff1a30e13f1deafd6c471f6b26594f023d3828ec23 not found: ID does not exist" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.642746 4696 scope.go:117] "RemoveContainer" containerID="473840c7910cf1720058db29c86564e69cb039cf4db0ba23a177368175ed1f58" Mar 21 08:33:26 crc kubenswrapper[4696]: E0321 08:33:26.642938 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"473840c7910cf1720058db29c86564e69cb039cf4db0ba23a177368175ed1f58\": container with ID starting with 473840c7910cf1720058db29c86564e69cb039cf4db0ba23a177368175ed1f58 not found: ID does not exist" containerID="473840c7910cf1720058db29c86564e69cb039cf4db0ba23a177368175ed1f58" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.642959 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"473840c7910cf1720058db29c86564e69cb039cf4db0ba23a177368175ed1f58"} err="failed to get container status \"473840c7910cf1720058db29c86564e69cb039cf4db0ba23a177368175ed1f58\": rpc error: code = NotFound desc = could not find container \"473840c7910cf1720058db29c86564e69cb039cf4db0ba23a177368175ed1f58\": container with ID starting with 473840c7910cf1720058db29c86564e69cb039cf4db0ba23a177368175ed1f58 not found: ID does not exist" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.646019 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d70a46f0-1e5d-4edb-8745-9b8a41b66241-kube-api-access-fjjfw" (OuterVolumeSpecName: "kube-api-access-fjjfw") pod "d70a46f0-1e5d-4edb-8745-9b8a41b66241" (UID: "d70a46f0-1e5d-4edb-8745-9b8a41b66241"). InnerVolumeSpecName "kube-api-access-fjjfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.674268 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d70a46f0-1e5d-4edb-8745-9b8a41b66241-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d70a46f0-1e5d-4edb-8745-9b8a41b66241" (UID: "d70a46f0-1e5d-4edb-8745-9b8a41b66241"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.715445 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5vrbx"] Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.722254 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5vrbx"] Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.736721 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v9rlg"] Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.740466 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v9rlg"] Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.742462 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d70a46f0-1e5d-4edb-8745-9b8a41b66241-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.742489 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d70a46f0-1e5d-4edb-8745-9b8a41b66241-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.742504 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjjfw\" (UniqueName: \"kubernetes.io/projected/d70a46f0-1e5d-4edb-8745-9b8a41b66241-kube-api-access-fjjfw\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.747872 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cfjs8"] Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.756133 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cfjs8"] Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.806826 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w6xbt"] Mar 21 08:33:26 crc kubenswrapper[4696]: W0321 08:33:26.811187 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd45283e8_56c7_4a2e_8a7e_88e459bb1c04.slice/crio-e219ad427515080d3c20a6a0439c2c0868104f19afb60146441294b59cc3d967 WatchSource:0}: Error finding container e219ad427515080d3c20a6a0439c2c0868104f19afb60146441294b59cc3d967: Status 404 returned error can't find the container with id e219ad427515080d3c20a6a0439c2c0868104f19afb60146441294b59cc3d967 Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.825465 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qsg67" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.953461 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gswrs\" (UniqueName: \"kubernetes.io/projected/6aabaca2-7def-4550-bae1-9337b7c1002b-kube-api-access-gswrs\") pod \"6aabaca2-7def-4550-bae1-9337b7c1002b\" (UID: \"6aabaca2-7def-4550-bae1-9337b7c1002b\") " Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.953543 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6aabaca2-7def-4550-bae1-9337b7c1002b-utilities\") pod \"6aabaca2-7def-4550-bae1-9337b7c1002b\" (UID: \"6aabaca2-7def-4550-bae1-9337b7c1002b\") " Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.953584 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6aabaca2-7def-4550-bae1-9337b7c1002b-catalog-content\") pod \"6aabaca2-7def-4550-bae1-9337b7c1002b\" (UID: \"6aabaca2-7def-4550-bae1-9337b7c1002b\") " Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.956548 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6aabaca2-7def-4550-bae1-9337b7c1002b-utilities" (OuterVolumeSpecName: "utilities") pod "6aabaca2-7def-4550-bae1-9337b7c1002b" (UID: "6aabaca2-7def-4550-bae1-9337b7c1002b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:33:26 crc kubenswrapper[4696]: I0321 08:33:26.960209 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6aabaca2-7def-4550-bae1-9337b7c1002b-kube-api-access-gswrs" (OuterVolumeSpecName: "kube-api-access-gswrs") pod "6aabaca2-7def-4550-bae1-9337b7c1002b" (UID: "6aabaca2-7def-4550-bae1-9337b7c1002b"). InnerVolumeSpecName "kube-api-access-gswrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.006508 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6aabaca2-7def-4550-bae1-9337b7c1002b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6aabaca2-7def-4550-bae1-9337b7c1002b" (UID: "6aabaca2-7def-4550-bae1-9337b7c1002b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.054750 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gswrs\" (UniqueName: \"kubernetes.io/projected/6aabaca2-7def-4550-bae1-9337b7c1002b-kube-api-access-gswrs\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.054782 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6aabaca2-7def-4550-bae1-9337b7c1002b-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.054792 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6aabaca2-7def-4550-bae1-9337b7c1002b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.431716 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jnw8m" event={"ID":"d70a46f0-1e5d-4edb-8745-9b8a41b66241","Type":"ContainerDied","Data":"c421f19fe2b512c1558d4402b703acf79f57365a6fbf5ec4ecb952e7c7904356"} Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.431762 4696 scope.go:117] "RemoveContainer" containerID="ad76337a74c84f6126a24fc0768cd55ff27a998a9d26ef3c78b7b12073b148c3" Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.432059 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jnw8m" Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.434877 4696 generic.go:334] "Generic (PLEG): container finished" podID="6aabaca2-7def-4550-bae1-9337b7c1002b" containerID="bdc0c304327cc59729598ff33c42a54e45d127e5d42d5558ef2977e434010ea4" exitCode=0 Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.434944 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qsg67" Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.434943 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qsg67" event={"ID":"6aabaca2-7def-4550-bae1-9337b7c1002b","Type":"ContainerDied","Data":"bdc0c304327cc59729598ff33c42a54e45d127e5d42d5558ef2977e434010ea4"} Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.435117 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qsg67" event={"ID":"6aabaca2-7def-4550-bae1-9337b7c1002b","Type":"ContainerDied","Data":"ab7fb5221a42f276201e0f172904818436768bf0492bc00ee90713be326befaf"} Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.436371 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-w6xbt" event={"ID":"d45283e8-56c7-4a2e-8a7e-88e459bb1c04","Type":"ContainerStarted","Data":"03ea44d10be9840b8f7d2249c26a275a9dfd61b42fedaf29fd9e521796c1838e"} Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.436426 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-w6xbt" event={"ID":"d45283e8-56c7-4a2e-8a7e-88e459bb1c04","Type":"ContainerStarted","Data":"e219ad427515080d3c20a6a0439c2c0868104f19afb60146441294b59cc3d967"} Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.436763 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-w6xbt" Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.441684 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-w6xbt" Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.449363 4696 scope.go:117] "RemoveContainer" containerID="d29438d9bfb377184180255d89e71c9df999fdb0b903dc06d7601c2fc7d6a1a3" Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.453809 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-w6xbt" podStartSLOduration=2.453791484 podStartE2EDuration="2.453791484s" podCreationTimestamp="2026-03-21 08:33:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:33:27.45128431 +0000 UTC m=+341.572165033" watchObservedRunningTime="2026-03-21 08:33:27.453791484 +0000 UTC m=+341.574672197" Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.472870 4696 scope.go:117] "RemoveContainer" containerID="8e3278a745c964032c3d9d433afe57aa71f2dcc8c63adac2d0008719bbfb7122" Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.475457 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jnw8m"] Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.482725 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jnw8m"] Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.497398 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qsg67"] Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.498214 4696 scope.go:117] "RemoveContainer" containerID="bdc0c304327cc59729598ff33c42a54e45d127e5d42d5558ef2977e434010ea4" Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.503417 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qsg67"] Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.526481 4696 scope.go:117] "RemoveContainer" containerID="c41c5c0054693270a82a25c88765741367992244e855001bd0c5b6a6fa6345a9" Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.538429 4696 scope.go:117] "RemoveContainer" containerID="82b676e04b7e68c50b5598983dadfa70abef0bed245e0d350cb58d9c32cc7f47" Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.559970 4696 scope.go:117] "RemoveContainer" containerID="bdc0c304327cc59729598ff33c42a54e45d127e5d42d5558ef2977e434010ea4" Mar 21 08:33:27 crc kubenswrapper[4696]: E0321 08:33:27.562147 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdc0c304327cc59729598ff33c42a54e45d127e5d42d5558ef2977e434010ea4\": container with ID starting with bdc0c304327cc59729598ff33c42a54e45d127e5d42d5558ef2977e434010ea4 not found: ID does not exist" containerID="bdc0c304327cc59729598ff33c42a54e45d127e5d42d5558ef2977e434010ea4" Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.562211 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdc0c304327cc59729598ff33c42a54e45d127e5d42d5558ef2977e434010ea4"} err="failed to get container status \"bdc0c304327cc59729598ff33c42a54e45d127e5d42d5558ef2977e434010ea4\": rpc error: code = NotFound desc = could not find container \"bdc0c304327cc59729598ff33c42a54e45d127e5d42d5558ef2977e434010ea4\": container with ID starting with bdc0c304327cc59729598ff33c42a54e45d127e5d42d5558ef2977e434010ea4 not found: ID does not exist" Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.562245 4696 scope.go:117] "RemoveContainer" containerID="c41c5c0054693270a82a25c88765741367992244e855001bd0c5b6a6fa6345a9" Mar 21 08:33:27 crc kubenswrapper[4696]: E0321 08:33:27.562738 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c41c5c0054693270a82a25c88765741367992244e855001bd0c5b6a6fa6345a9\": container with ID starting with c41c5c0054693270a82a25c88765741367992244e855001bd0c5b6a6fa6345a9 not found: ID does not exist" containerID="c41c5c0054693270a82a25c88765741367992244e855001bd0c5b6a6fa6345a9" Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.562782 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c41c5c0054693270a82a25c88765741367992244e855001bd0c5b6a6fa6345a9"} err="failed to get container status \"c41c5c0054693270a82a25c88765741367992244e855001bd0c5b6a6fa6345a9\": rpc error: code = NotFound desc = could not find container \"c41c5c0054693270a82a25c88765741367992244e855001bd0c5b6a6fa6345a9\": container with ID starting with c41c5c0054693270a82a25c88765741367992244e855001bd0c5b6a6fa6345a9 not found: ID does not exist" Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.562798 4696 scope.go:117] "RemoveContainer" containerID="82b676e04b7e68c50b5598983dadfa70abef0bed245e0d350cb58d9c32cc7f47" Mar 21 08:33:27 crc kubenswrapper[4696]: E0321 08:33:27.563101 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82b676e04b7e68c50b5598983dadfa70abef0bed245e0d350cb58d9c32cc7f47\": container with ID starting with 82b676e04b7e68c50b5598983dadfa70abef0bed245e0d350cb58d9c32cc7f47 not found: ID does not exist" containerID="82b676e04b7e68c50b5598983dadfa70abef0bed245e0d350cb58d9c32cc7f47" Mar 21 08:33:27 crc kubenswrapper[4696]: I0321 08:33:27.563138 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82b676e04b7e68c50b5598983dadfa70abef0bed245e0d350cb58d9c32cc7f47"} err="failed to get container status \"82b676e04b7e68c50b5598983dadfa70abef0bed245e0d350cb58d9c32cc7f47\": rpc error: code = NotFound desc = could not find container \"82b676e04b7e68c50b5598983dadfa70abef0bed245e0d350cb58d9c32cc7f47\": container with ID starting with 82b676e04b7e68c50b5598983dadfa70abef0bed245e0d350cb58d9c32cc7f47 not found: ID does not exist" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.096185 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mmkz6"] Mar 21 08:33:28 crc kubenswrapper[4696]: E0321 08:33:28.096922 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d70a46f0-1e5d-4edb-8745-9b8a41b66241" containerName="extract-content" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.096940 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d70a46f0-1e5d-4edb-8745-9b8a41b66241" containerName="extract-content" Mar 21 08:33:28 crc kubenswrapper[4696]: E0321 08:33:28.096960 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85a5000b-3b68-4808-bcfc-93bcbde10dde" containerName="registry-server" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.096969 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="85a5000b-3b68-4808-bcfc-93bcbde10dde" containerName="registry-server" Mar 21 08:33:28 crc kubenswrapper[4696]: E0321 08:33:28.096983 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85a5000b-3b68-4808-bcfc-93bcbde10dde" containerName="extract-content" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.096992 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="85a5000b-3b68-4808-bcfc-93bcbde10dde" containerName="extract-content" Mar 21 08:33:28 crc kubenswrapper[4696]: E0321 08:33:28.097002 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d70a46f0-1e5d-4edb-8745-9b8a41b66241" containerName="extract-utilities" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.097011 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d70a46f0-1e5d-4edb-8745-9b8a41b66241" containerName="extract-utilities" Mar 21 08:33:28 crc kubenswrapper[4696]: E0321 08:33:28.097023 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0030cc8a-f473-4d0a-ace9-728b66e62634" containerName="registry-server" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.097031 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0030cc8a-f473-4d0a-ace9-728b66e62634" containerName="registry-server" Mar 21 08:33:28 crc kubenswrapper[4696]: E0321 08:33:28.097042 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0030cc8a-f473-4d0a-ace9-728b66e62634" containerName="extract-content" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.097052 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0030cc8a-f473-4d0a-ace9-728b66e62634" containerName="extract-content" Mar 21 08:33:28 crc kubenswrapper[4696]: E0321 08:33:28.097061 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aabaca2-7def-4550-bae1-9337b7c1002b" containerName="extract-content" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.097069 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aabaca2-7def-4550-bae1-9337b7c1002b" containerName="extract-content" Mar 21 08:33:28 crc kubenswrapper[4696]: E0321 08:33:28.097079 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0030cc8a-f473-4d0a-ace9-728b66e62634" containerName="extract-utilities" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.097087 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0030cc8a-f473-4d0a-ace9-728b66e62634" containerName="extract-utilities" Mar 21 08:33:28 crc kubenswrapper[4696]: E0321 08:33:28.097098 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="973fa2b5-c6ce-498f-9b2b-311e8dd34057" containerName="marketplace-operator" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.097106 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="973fa2b5-c6ce-498f-9b2b-311e8dd34057" containerName="marketplace-operator" Mar 21 08:33:28 crc kubenswrapper[4696]: E0321 08:33:28.097118 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d70a46f0-1e5d-4edb-8745-9b8a41b66241" containerName="registry-server" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.097126 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d70a46f0-1e5d-4edb-8745-9b8a41b66241" containerName="registry-server" Mar 21 08:33:28 crc kubenswrapper[4696]: E0321 08:33:28.097135 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85a5000b-3b68-4808-bcfc-93bcbde10dde" containerName="extract-utilities" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.097142 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="85a5000b-3b68-4808-bcfc-93bcbde10dde" containerName="extract-utilities" Mar 21 08:33:28 crc kubenswrapper[4696]: E0321 08:33:28.097156 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aabaca2-7def-4550-bae1-9337b7c1002b" containerName="extract-utilities" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.097164 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aabaca2-7def-4550-bae1-9337b7c1002b" containerName="extract-utilities" Mar 21 08:33:28 crc kubenswrapper[4696]: E0321 08:33:28.097175 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aabaca2-7def-4550-bae1-9337b7c1002b" containerName="registry-server" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.097184 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aabaca2-7def-4550-bae1-9337b7c1002b" containerName="registry-server" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.097299 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="973fa2b5-c6ce-498f-9b2b-311e8dd34057" containerName="marketplace-operator" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.097315 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="d70a46f0-1e5d-4edb-8745-9b8a41b66241" containerName="registry-server" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.097327 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="0030cc8a-f473-4d0a-ace9-728b66e62634" containerName="registry-server" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.097342 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aabaca2-7def-4550-bae1-9337b7c1002b" containerName="registry-server" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.097356 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="85a5000b-3b68-4808-bcfc-93bcbde10dde" containerName="registry-server" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.098259 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mmkz6" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.102353 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.105365 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mmkz6"] Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.168296 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77-utilities\") pod \"redhat-marketplace-mmkz6\" (UID: \"5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77\") " pod="openshift-marketplace/redhat-marketplace-mmkz6" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.168394 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cljdw\" (UniqueName: \"kubernetes.io/projected/5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77-kube-api-access-cljdw\") pod \"redhat-marketplace-mmkz6\" (UID: \"5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77\") " pod="openshift-marketplace/redhat-marketplace-mmkz6" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.168584 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77-catalog-content\") pod \"redhat-marketplace-mmkz6\" (UID: \"5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77\") " pod="openshift-marketplace/redhat-marketplace-mmkz6" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.270117 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cljdw\" (UniqueName: \"kubernetes.io/projected/5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77-kube-api-access-cljdw\") pod \"redhat-marketplace-mmkz6\" (UID: \"5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77\") " pod="openshift-marketplace/redhat-marketplace-mmkz6" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.270425 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77-catalog-content\") pod \"redhat-marketplace-mmkz6\" (UID: \"5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77\") " pod="openshift-marketplace/redhat-marketplace-mmkz6" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.270895 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77-catalog-content\") pod \"redhat-marketplace-mmkz6\" (UID: \"5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77\") " pod="openshift-marketplace/redhat-marketplace-mmkz6" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.270911 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77-utilities\") pod \"redhat-marketplace-mmkz6\" (UID: \"5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77\") " pod="openshift-marketplace/redhat-marketplace-mmkz6" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.271174 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77-utilities\") pod \"redhat-marketplace-mmkz6\" (UID: \"5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77\") " pod="openshift-marketplace/redhat-marketplace-mmkz6" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.306090 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cljdw\" (UniqueName: \"kubernetes.io/projected/5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77-kube-api-access-cljdw\") pod \"redhat-marketplace-mmkz6\" (UID: \"5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77\") " pod="openshift-marketplace/redhat-marketplace-mmkz6" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.315435 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n6f4n"] Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.317671 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n6f4n" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.319831 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.322107 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n6f4n"] Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.372483 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-797lp\" (UniqueName: \"kubernetes.io/projected/b49f2cd9-95cf-4749-889e-c47d027c253b-kube-api-access-797lp\") pod \"redhat-operators-n6f4n\" (UID: \"b49f2cd9-95cf-4749-889e-c47d027c253b\") " pod="openshift-marketplace/redhat-operators-n6f4n" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.372715 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b49f2cd9-95cf-4749-889e-c47d027c253b-catalog-content\") pod \"redhat-operators-n6f4n\" (UID: \"b49f2cd9-95cf-4749-889e-c47d027c253b\") " pod="openshift-marketplace/redhat-operators-n6f4n" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.372805 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b49f2cd9-95cf-4749-889e-c47d027c253b-utilities\") pod \"redhat-operators-n6f4n\" (UID: \"b49f2cd9-95cf-4749-889e-c47d027c253b\") " pod="openshift-marketplace/redhat-operators-n6f4n" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.417890 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mmkz6" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.474624 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-797lp\" (UniqueName: \"kubernetes.io/projected/b49f2cd9-95cf-4749-889e-c47d027c253b-kube-api-access-797lp\") pod \"redhat-operators-n6f4n\" (UID: \"b49f2cd9-95cf-4749-889e-c47d027c253b\") " pod="openshift-marketplace/redhat-operators-n6f4n" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.474780 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b49f2cd9-95cf-4749-889e-c47d027c253b-catalog-content\") pod \"redhat-operators-n6f4n\" (UID: \"b49f2cd9-95cf-4749-889e-c47d027c253b\") " pod="openshift-marketplace/redhat-operators-n6f4n" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.474899 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b49f2cd9-95cf-4749-889e-c47d027c253b-utilities\") pod \"redhat-operators-n6f4n\" (UID: \"b49f2cd9-95cf-4749-889e-c47d027c253b\") " pod="openshift-marketplace/redhat-operators-n6f4n" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.475608 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b49f2cd9-95cf-4749-889e-c47d027c253b-utilities\") pod \"redhat-operators-n6f4n\" (UID: \"b49f2cd9-95cf-4749-889e-c47d027c253b\") " pod="openshift-marketplace/redhat-operators-n6f4n" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.477947 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b49f2cd9-95cf-4749-889e-c47d027c253b-catalog-content\") pod \"redhat-operators-n6f4n\" (UID: \"b49f2cd9-95cf-4749-889e-c47d027c253b\") " pod="openshift-marketplace/redhat-operators-n6f4n" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.497456 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-797lp\" (UniqueName: \"kubernetes.io/projected/b49f2cd9-95cf-4749-889e-c47d027c253b-kube-api-access-797lp\") pod \"redhat-operators-n6f4n\" (UID: \"b49f2cd9-95cf-4749-889e-c47d027c253b\") " pod="openshift-marketplace/redhat-operators-n6f4n" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.546078 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0030cc8a-f473-4d0a-ace9-728b66e62634" path="/var/lib/kubelet/pods/0030cc8a-f473-4d0a-ace9-728b66e62634/volumes" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.548271 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6aabaca2-7def-4550-bae1-9337b7c1002b" path="/var/lib/kubelet/pods/6aabaca2-7def-4550-bae1-9337b7c1002b/volumes" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.549661 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85a5000b-3b68-4808-bcfc-93bcbde10dde" path="/var/lib/kubelet/pods/85a5000b-3b68-4808-bcfc-93bcbde10dde/volumes" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.552121 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="973fa2b5-c6ce-498f-9b2b-311e8dd34057" path="/var/lib/kubelet/pods/973fa2b5-c6ce-498f-9b2b-311e8dd34057/volumes" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.553203 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d70a46f0-1e5d-4edb-8745-9b8a41b66241" path="/var/lib/kubelet/pods/d70a46f0-1e5d-4edb-8745-9b8a41b66241/volumes" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.666329 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n6f4n" Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.682447 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mmkz6"] Mar 21 08:33:28 crc kubenswrapper[4696]: W0321 08:33:28.684697 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cb0e56a_eac9_4103_8d0c_e5ebc1c5dd77.slice/crio-7df304284cd48615f5b6d0a3dfc559a86804646090216e701dfcc173b98e3142 WatchSource:0}: Error finding container 7df304284cd48615f5b6d0a3dfc559a86804646090216e701dfcc173b98e3142: Status 404 returned error can't find the container with id 7df304284cd48615f5b6d0a3dfc559a86804646090216e701dfcc173b98e3142 Mar 21 08:33:28 crc kubenswrapper[4696]: I0321 08:33:28.866251 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n6f4n"] Mar 21 08:33:28 crc kubenswrapper[4696]: W0321 08:33:28.884474 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb49f2cd9_95cf_4749_889e_c47d027c253b.slice/crio-1391ad5b112148b56394953ff0c7e233922a389753a4c3f538f4d2e88bd09446 WatchSource:0}: Error finding container 1391ad5b112148b56394953ff0c7e233922a389753a4c3f538f4d2e88bd09446: Status 404 returned error can't find the container with id 1391ad5b112148b56394953ff0c7e233922a389753a4c3f538f4d2e88bd09446 Mar 21 08:33:29 crc kubenswrapper[4696]: I0321 08:33:29.462721 4696 generic.go:334] "Generic (PLEG): container finished" podID="5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77" containerID="04f6ced3dfc877d8402ff3c0090358383d5c1e2b2ca69bd264bd349e281f05ab" exitCode=0 Mar 21 08:33:29 crc kubenswrapper[4696]: I0321 08:33:29.462843 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mmkz6" event={"ID":"5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77","Type":"ContainerDied","Data":"04f6ced3dfc877d8402ff3c0090358383d5c1e2b2ca69bd264bd349e281f05ab"} Mar 21 08:33:29 crc kubenswrapper[4696]: I0321 08:33:29.462934 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mmkz6" event={"ID":"5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77","Type":"ContainerStarted","Data":"7df304284cd48615f5b6d0a3dfc559a86804646090216e701dfcc173b98e3142"} Mar 21 08:33:29 crc kubenswrapper[4696]: I0321 08:33:29.464338 4696 generic.go:334] "Generic (PLEG): container finished" podID="b49f2cd9-95cf-4749-889e-c47d027c253b" containerID="0d5c4a1e6f8014194291ea9a9fbc94110130a7b3f04f32146b897aec6c981f32" exitCode=0 Mar 21 08:33:29 crc kubenswrapper[4696]: I0321 08:33:29.464372 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6f4n" event={"ID":"b49f2cd9-95cf-4749-889e-c47d027c253b","Type":"ContainerDied","Data":"0d5c4a1e6f8014194291ea9a9fbc94110130a7b3f04f32146b897aec6c981f32"} Mar 21 08:33:29 crc kubenswrapper[4696]: I0321 08:33:29.464397 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6f4n" event={"ID":"b49f2cd9-95cf-4749-889e-c47d027c253b","Type":"ContainerStarted","Data":"1391ad5b112148b56394953ff0c7e233922a389753a4c3f538f4d2e88bd09446"} Mar 21 08:33:30 crc kubenswrapper[4696]: E0321 08:33:30.337904 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cb0e56a_eac9_4103_8d0c_e5ebc1c5dd77.slice/crio-9f7054cd839c9f502efecc79ff1b13209ac331a3566074a474735499788c60a2.scope\": RecentStats: unable to find data in memory cache]" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.471850 4696 generic.go:334] "Generic (PLEG): container finished" podID="5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77" containerID="9f7054cd839c9f502efecc79ff1b13209ac331a3566074a474735499788c60a2" exitCode=0 Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.471949 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mmkz6" event={"ID":"5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77","Type":"ContainerDied","Data":"9f7054cd839c9f502efecc79ff1b13209ac331a3566074a474735499788c60a2"} Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.474981 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6f4n" event={"ID":"b49f2cd9-95cf-4749-889e-c47d027c253b","Type":"ContainerStarted","Data":"c204273af4b2483c25c0153d61c2479609a94a2a7cb6528379cfbbb3e6c34fe8"} Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.494331 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rwzg5"] Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.495319 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rwzg5" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.497992 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.511516 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rwzg5"] Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.602977 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqr6d\" (UniqueName: \"kubernetes.io/projected/1050b6cc-b45d-42cb-8d54-7104988c8225-kube-api-access-rqr6d\") pod \"community-operators-rwzg5\" (UID: \"1050b6cc-b45d-42cb-8d54-7104988c8225\") " pod="openshift-marketplace/community-operators-rwzg5" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.603177 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1050b6cc-b45d-42cb-8d54-7104988c8225-catalog-content\") pod \"community-operators-rwzg5\" (UID: \"1050b6cc-b45d-42cb-8d54-7104988c8225\") " pod="openshift-marketplace/community-operators-rwzg5" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.603900 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1050b6cc-b45d-42cb-8d54-7104988c8225-utilities\") pod \"community-operators-rwzg5\" (UID: \"1050b6cc-b45d-42cb-8d54-7104988c8225\") " pod="openshift-marketplace/community-operators-rwzg5" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.690713 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-g5vlg"] Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.691585 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g5vlg" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.693184 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.700071 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g5vlg"] Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.704647 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1050b6cc-b45d-42cb-8d54-7104988c8225-catalog-content\") pod \"community-operators-rwzg5\" (UID: \"1050b6cc-b45d-42cb-8d54-7104988c8225\") " pod="openshift-marketplace/community-operators-rwzg5" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.704683 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1050b6cc-b45d-42cb-8d54-7104988c8225-utilities\") pod \"community-operators-rwzg5\" (UID: \"1050b6cc-b45d-42cb-8d54-7104988c8225\") " pod="openshift-marketplace/community-operators-rwzg5" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.704716 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqr6d\" (UniqueName: \"kubernetes.io/projected/1050b6cc-b45d-42cb-8d54-7104988c8225-kube-api-access-rqr6d\") pod \"community-operators-rwzg5\" (UID: \"1050b6cc-b45d-42cb-8d54-7104988c8225\") " pod="openshift-marketplace/community-operators-rwzg5" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.705066 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1050b6cc-b45d-42cb-8d54-7104988c8225-catalog-content\") pod \"community-operators-rwzg5\" (UID: \"1050b6cc-b45d-42cb-8d54-7104988c8225\") " pod="openshift-marketplace/community-operators-rwzg5" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.705195 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1050b6cc-b45d-42cb-8d54-7104988c8225-utilities\") pod \"community-operators-rwzg5\" (UID: \"1050b6cc-b45d-42cb-8d54-7104988c8225\") " pod="openshift-marketplace/community-operators-rwzg5" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.724433 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqr6d\" (UniqueName: \"kubernetes.io/projected/1050b6cc-b45d-42cb-8d54-7104988c8225-kube-api-access-rqr6d\") pod \"community-operators-rwzg5\" (UID: \"1050b6cc-b45d-42cb-8d54-7104988c8225\") " pod="openshift-marketplace/community-operators-rwzg5" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.805865 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e175ab46-d3f4-4051-a86e-dcf2727b9efd-utilities\") pod \"certified-operators-g5vlg\" (UID: \"e175ab46-d3f4-4051-a86e-dcf2727b9efd\") " pod="openshift-marketplace/certified-operators-g5vlg" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.805912 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e175ab46-d3f4-4051-a86e-dcf2727b9efd-catalog-content\") pod \"certified-operators-g5vlg\" (UID: \"e175ab46-d3f4-4051-a86e-dcf2727b9efd\") " pod="openshift-marketplace/certified-operators-g5vlg" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.805944 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkfj8\" (UniqueName: \"kubernetes.io/projected/e175ab46-d3f4-4051-a86e-dcf2727b9efd-kube-api-access-gkfj8\") pod \"certified-operators-g5vlg\" (UID: \"e175ab46-d3f4-4051-a86e-dcf2727b9efd\") " pod="openshift-marketplace/certified-operators-g5vlg" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.809584 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rwzg5" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.907470 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e175ab46-d3f4-4051-a86e-dcf2727b9efd-utilities\") pod \"certified-operators-g5vlg\" (UID: \"e175ab46-d3f4-4051-a86e-dcf2727b9efd\") " pod="openshift-marketplace/certified-operators-g5vlg" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.907508 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e175ab46-d3f4-4051-a86e-dcf2727b9efd-catalog-content\") pod \"certified-operators-g5vlg\" (UID: \"e175ab46-d3f4-4051-a86e-dcf2727b9efd\") " pod="openshift-marketplace/certified-operators-g5vlg" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.907539 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkfj8\" (UniqueName: \"kubernetes.io/projected/e175ab46-d3f4-4051-a86e-dcf2727b9efd-kube-api-access-gkfj8\") pod \"certified-operators-g5vlg\" (UID: \"e175ab46-d3f4-4051-a86e-dcf2727b9efd\") " pod="openshift-marketplace/certified-operators-g5vlg" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.908065 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e175ab46-d3f4-4051-a86e-dcf2727b9efd-utilities\") pod \"certified-operators-g5vlg\" (UID: \"e175ab46-d3f4-4051-a86e-dcf2727b9efd\") " pod="openshift-marketplace/certified-operators-g5vlg" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.908032 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e175ab46-d3f4-4051-a86e-dcf2727b9efd-catalog-content\") pod \"certified-operators-g5vlg\" (UID: \"e175ab46-d3f4-4051-a86e-dcf2727b9efd\") " pod="openshift-marketplace/certified-operators-g5vlg" Mar 21 08:33:30 crc kubenswrapper[4696]: I0321 08:33:30.926290 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkfj8\" (UniqueName: \"kubernetes.io/projected/e175ab46-d3f4-4051-a86e-dcf2727b9efd-kube-api-access-gkfj8\") pod \"certified-operators-g5vlg\" (UID: \"e175ab46-d3f4-4051-a86e-dcf2727b9efd\") " pod="openshift-marketplace/certified-operators-g5vlg" Mar 21 08:33:31 crc kubenswrapper[4696]: I0321 08:33:31.069978 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g5vlg" Mar 21 08:33:31 crc kubenswrapper[4696]: I0321 08:33:31.214613 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rwzg5"] Mar 21 08:33:31 crc kubenswrapper[4696]: I0321 08:33:31.455023 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g5vlg"] Mar 21 08:33:31 crc kubenswrapper[4696]: W0321 08:33:31.455168 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode175ab46_d3f4_4051_a86e_dcf2727b9efd.slice/crio-5019f4219fffeb4292a5375bcb3fcf6a1b2d0d3b803466a0c36c44126246c934 WatchSource:0}: Error finding container 5019f4219fffeb4292a5375bcb3fcf6a1b2d0d3b803466a0c36c44126246c934: Status 404 returned error can't find the container with id 5019f4219fffeb4292a5375bcb3fcf6a1b2d0d3b803466a0c36c44126246c934 Mar 21 08:33:31 crc kubenswrapper[4696]: I0321 08:33:31.485807 4696 generic.go:334] "Generic (PLEG): container finished" podID="b49f2cd9-95cf-4749-889e-c47d027c253b" containerID="c204273af4b2483c25c0153d61c2479609a94a2a7cb6528379cfbbb3e6c34fe8" exitCode=0 Mar 21 08:33:31 crc kubenswrapper[4696]: I0321 08:33:31.486574 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6f4n" event={"ID":"b49f2cd9-95cf-4749-889e-c47d027c253b","Type":"ContainerDied","Data":"c204273af4b2483c25c0153d61c2479609a94a2a7cb6528379cfbbb3e6c34fe8"} Mar 21 08:33:31 crc kubenswrapper[4696]: I0321 08:33:31.487568 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5vlg" event={"ID":"e175ab46-d3f4-4051-a86e-dcf2727b9efd","Type":"ContainerStarted","Data":"5019f4219fffeb4292a5375bcb3fcf6a1b2d0d3b803466a0c36c44126246c934"} Mar 21 08:33:31 crc kubenswrapper[4696]: I0321 08:33:31.488956 4696 generic.go:334] "Generic (PLEG): container finished" podID="1050b6cc-b45d-42cb-8d54-7104988c8225" containerID="793b9f73b6105a80904237ca1ccf096e03b849e693a74816505aeef5ca85969c" exitCode=0 Mar 21 08:33:31 crc kubenswrapper[4696]: I0321 08:33:31.489027 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwzg5" event={"ID":"1050b6cc-b45d-42cb-8d54-7104988c8225","Type":"ContainerDied","Data":"793b9f73b6105a80904237ca1ccf096e03b849e693a74816505aeef5ca85969c"} Mar 21 08:33:31 crc kubenswrapper[4696]: I0321 08:33:31.489070 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwzg5" event={"ID":"1050b6cc-b45d-42cb-8d54-7104988c8225","Type":"ContainerStarted","Data":"813b3f3a5cf6ed6a61f7cf801d896fd706b11690a5455132aed3033b8c9bfd52"} Mar 21 08:33:31 crc kubenswrapper[4696]: I0321 08:33:31.493519 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mmkz6" event={"ID":"5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77","Type":"ContainerStarted","Data":"6fd36c9dfe603dd313bd8a07c8af6ef6b7e5cc50f1c645746baf5c2b7498920c"} Mar 21 08:33:31 crc kubenswrapper[4696]: I0321 08:33:31.560066 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mmkz6" podStartSLOduration=2.113497117 podStartE2EDuration="3.560049333s" podCreationTimestamp="2026-03-21 08:33:28 +0000 UTC" firstStartedPulling="2026-03-21 08:33:29.468024492 +0000 UTC m=+343.588905245" lastFinishedPulling="2026-03-21 08:33:30.914576748 +0000 UTC m=+345.035457461" observedRunningTime="2026-03-21 08:33:31.540235874 +0000 UTC m=+345.661116597" watchObservedRunningTime="2026-03-21 08:33:31.560049333 +0000 UTC m=+345.680930046" Mar 21 08:33:32 crc kubenswrapper[4696]: I0321 08:33:32.500759 4696 generic.go:334] "Generic (PLEG): container finished" podID="1050b6cc-b45d-42cb-8d54-7104988c8225" containerID="f9d78988a08976f1ae122859835666faefc86f66c882c0f63288eec7d7f10db4" exitCode=0 Mar 21 08:33:32 crc kubenswrapper[4696]: I0321 08:33:32.500966 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwzg5" event={"ID":"1050b6cc-b45d-42cb-8d54-7104988c8225","Type":"ContainerDied","Data":"f9d78988a08976f1ae122859835666faefc86f66c882c0f63288eec7d7f10db4"} Mar 21 08:33:32 crc kubenswrapper[4696]: I0321 08:33:32.504386 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6f4n" event={"ID":"b49f2cd9-95cf-4749-889e-c47d027c253b","Type":"ContainerStarted","Data":"bb3d2be94f2e404ecbb0a5130c4726afa49a5df09c407dda94f7d55e81d8acc1"} Mar 21 08:33:32 crc kubenswrapper[4696]: I0321 08:33:32.507158 4696 generic.go:334] "Generic (PLEG): container finished" podID="e175ab46-d3f4-4051-a86e-dcf2727b9efd" containerID="de5149f1aa50f42ceb7405f3a725a7f1958ae0365efb197eaf4365cdf60a2320" exitCode=0 Mar 21 08:33:32 crc kubenswrapper[4696]: I0321 08:33:32.508102 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5vlg" event={"ID":"e175ab46-d3f4-4051-a86e-dcf2727b9efd","Type":"ContainerDied","Data":"de5149f1aa50f42ceb7405f3a725a7f1958ae0365efb197eaf4365cdf60a2320"} Mar 21 08:33:32 crc kubenswrapper[4696]: I0321 08:33:32.563500 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n6f4n" podStartSLOduration=1.927728358 podStartE2EDuration="4.56348231s" podCreationTimestamp="2026-03-21 08:33:28 +0000 UTC" firstStartedPulling="2026-03-21 08:33:29.467991531 +0000 UTC m=+343.588872244" lastFinishedPulling="2026-03-21 08:33:32.103745483 +0000 UTC m=+346.224626196" observedRunningTime="2026-03-21 08:33:32.559180864 +0000 UTC m=+346.680061587" watchObservedRunningTime="2026-03-21 08:33:32.56348231 +0000 UTC m=+346.684363023" Mar 21 08:33:33 crc kubenswrapper[4696]: I0321 08:33:33.189837 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-fkf7p" Mar 21 08:33:33 crc kubenswrapper[4696]: I0321 08:33:33.257309 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c46nr"] Mar 21 08:33:33 crc kubenswrapper[4696]: I0321 08:33:33.514781 4696 generic.go:334] "Generic (PLEG): container finished" podID="e175ab46-d3f4-4051-a86e-dcf2727b9efd" containerID="ecc2dc0a0ee8d192cf90ba5bbdb470bc5272e21ac5b9724252f741482c3309b6" exitCode=0 Mar 21 08:33:33 crc kubenswrapper[4696]: I0321 08:33:33.515126 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5vlg" event={"ID":"e175ab46-d3f4-4051-a86e-dcf2727b9efd","Type":"ContainerDied","Data":"ecc2dc0a0ee8d192cf90ba5bbdb470bc5272e21ac5b9724252f741482c3309b6"} Mar 21 08:33:33 crc kubenswrapper[4696]: I0321 08:33:33.523092 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwzg5" event={"ID":"1050b6cc-b45d-42cb-8d54-7104988c8225","Type":"ContainerStarted","Data":"a8e91362e31c15fe6d738afaa1b3eb961dc3cea15ec3d40e973a479fda9a126a"} Mar 21 08:33:33 crc kubenswrapper[4696]: I0321 08:33:33.562259 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rwzg5" podStartSLOduration=2.136387777 podStartE2EDuration="3.562243819s" podCreationTimestamp="2026-03-21 08:33:30 +0000 UTC" firstStartedPulling="2026-03-21 08:33:31.490303064 +0000 UTC m=+345.611183777" lastFinishedPulling="2026-03-21 08:33:32.916159116 +0000 UTC m=+347.037039819" observedRunningTime="2026-03-21 08:33:33.559691714 +0000 UTC m=+347.680572427" watchObservedRunningTime="2026-03-21 08:33:33.562243819 +0000 UTC m=+347.683124532" Mar 21 08:33:34 crc kubenswrapper[4696]: I0321 08:33:34.543685 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5vlg" event={"ID":"e175ab46-d3f4-4051-a86e-dcf2727b9efd","Type":"ContainerStarted","Data":"ec48350905de51df58691de852e1cb970729509ae705275407f08ea3c9a502f8"} Mar 21 08:33:34 crc kubenswrapper[4696]: I0321 08:33:34.561526 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-g5vlg" podStartSLOduration=3.109374545 podStartE2EDuration="4.561507354s" podCreationTimestamp="2026-03-21 08:33:30 +0000 UTC" firstStartedPulling="2026-03-21 08:33:32.50872886 +0000 UTC m=+346.629609573" lastFinishedPulling="2026-03-21 08:33:33.960861639 +0000 UTC m=+348.081742382" observedRunningTime="2026-03-21 08:33:34.55727596 +0000 UTC m=+348.678156693" watchObservedRunningTime="2026-03-21 08:33:34.561507354 +0000 UTC m=+348.682388077" Mar 21 08:33:38 crc kubenswrapper[4696]: I0321 08:33:38.419230 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mmkz6" Mar 21 08:33:38 crc kubenswrapper[4696]: I0321 08:33:38.420986 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mmkz6" Mar 21 08:33:38 crc kubenswrapper[4696]: I0321 08:33:38.472484 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mmkz6" Mar 21 08:33:38 crc kubenswrapper[4696]: I0321 08:33:38.622009 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mmkz6" Mar 21 08:33:38 crc kubenswrapper[4696]: I0321 08:33:38.666610 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n6f4n" Mar 21 08:33:38 crc kubenswrapper[4696]: I0321 08:33:38.666700 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n6f4n" Mar 21 08:33:39 crc kubenswrapper[4696]: I0321 08:33:39.711613 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n6f4n" podUID="b49f2cd9-95cf-4749-889e-c47d027c253b" containerName="registry-server" probeResult="failure" output=< Mar 21 08:33:39 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Mar 21 08:33:39 crc kubenswrapper[4696]: > Mar 21 08:33:40 crc kubenswrapper[4696]: I0321 08:33:40.810707 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rwzg5" Mar 21 08:33:40 crc kubenswrapper[4696]: I0321 08:33:40.811209 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rwzg5" Mar 21 08:33:40 crc kubenswrapper[4696]: I0321 08:33:40.880289 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rwzg5" Mar 21 08:33:41 crc kubenswrapper[4696]: I0321 08:33:41.070364 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-g5vlg" Mar 21 08:33:41 crc kubenswrapper[4696]: I0321 08:33:41.070462 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-g5vlg" Mar 21 08:33:41 crc kubenswrapper[4696]: I0321 08:33:41.139500 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-g5vlg" Mar 21 08:33:41 crc kubenswrapper[4696]: I0321 08:33:41.658681 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rwzg5" Mar 21 08:33:41 crc kubenswrapper[4696]: I0321 08:33:41.665752 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-g5vlg" Mar 21 08:33:48 crc kubenswrapper[4696]: I0321 08:33:48.732255 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n6f4n" Mar 21 08:33:48 crc kubenswrapper[4696]: I0321 08:33:48.811861 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n6f4n" Mar 21 08:33:58 crc kubenswrapper[4696]: I0321 08:33:58.301102 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" podUID="96b62643-30a8-47d5-8fe3-20eadf06e0f7" containerName="registry" containerID="cri-o://604fb7efc7658349c2c84a277fbbb4ba36f5ca4ba8e34e60ccebde2f771fc479" gracePeriod=30 Mar 21 08:33:58 crc kubenswrapper[4696]: I0321 08:33:58.701372 4696 generic.go:334] "Generic (PLEG): container finished" podID="96b62643-30a8-47d5-8fe3-20eadf06e0f7" containerID="604fb7efc7658349c2c84a277fbbb4ba36f5ca4ba8e34e60ccebde2f771fc479" exitCode=0 Mar 21 08:33:58 crc kubenswrapper[4696]: I0321 08:33:58.701422 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" event={"ID":"96b62643-30a8-47d5-8fe3-20eadf06e0f7","Type":"ContainerDied","Data":"604fb7efc7658349c2c84a277fbbb4ba36f5ca4ba8e34e60ccebde2f771fc479"} Mar 21 08:33:58 crc kubenswrapper[4696]: I0321 08:33:58.701716 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" event={"ID":"96b62643-30a8-47d5-8fe3-20eadf06e0f7","Type":"ContainerDied","Data":"f8fa3fbd51c5a21675e4119072861913eca893ce0c94a48c6e8092e25374c888"} Mar 21 08:33:58 crc kubenswrapper[4696]: I0321 08:33:58.701731 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8fa3fbd51c5a21675e4119072861913eca893ce0c94a48c6e8092e25374c888" Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.064696 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.253674 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/96b62643-30a8-47d5-8fe3-20eadf06e0f7-registry-tls\") pod \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.253952 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/96b62643-30a8-47d5-8fe3-20eadf06e0f7-installation-pull-secrets\") pod \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.254124 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.254172 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/96b62643-30a8-47d5-8fe3-20eadf06e0f7-ca-trust-extracted\") pod \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.254209 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/96b62643-30a8-47d5-8fe3-20eadf06e0f7-bound-sa-token\") pod \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.254234 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj4n5\" (UniqueName: \"kubernetes.io/projected/96b62643-30a8-47d5-8fe3-20eadf06e0f7-kube-api-access-pj4n5\") pod \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.254260 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96b62643-30a8-47d5-8fe3-20eadf06e0f7-trusted-ca\") pod \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.254279 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/96b62643-30a8-47d5-8fe3-20eadf06e0f7-registry-certificates\") pod \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\" (UID: \"96b62643-30a8-47d5-8fe3-20eadf06e0f7\") " Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.255388 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96b62643-30a8-47d5-8fe3-20eadf06e0f7-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "96b62643-30a8-47d5-8fe3-20eadf06e0f7" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.255603 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96b62643-30a8-47d5-8fe3-20eadf06e0f7-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "96b62643-30a8-47d5-8fe3-20eadf06e0f7" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.260664 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b62643-30a8-47d5-8fe3-20eadf06e0f7-kube-api-access-pj4n5" (OuterVolumeSpecName: "kube-api-access-pj4n5") pod "96b62643-30a8-47d5-8fe3-20eadf06e0f7" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7"). InnerVolumeSpecName "kube-api-access-pj4n5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.261634 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b62643-30a8-47d5-8fe3-20eadf06e0f7-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "96b62643-30a8-47d5-8fe3-20eadf06e0f7" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.263374 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b62643-30a8-47d5-8fe3-20eadf06e0f7-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "96b62643-30a8-47d5-8fe3-20eadf06e0f7" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.263665 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "96b62643-30a8-47d5-8fe3-20eadf06e0f7" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.277324 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b62643-30a8-47d5-8fe3-20eadf06e0f7-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "96b62643-30a8-47d5-8fe3-20eadf06e0f7" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.281857 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96b62643-30a8-47d5-8fe3-20eadf06e0f7-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "96b62643-30a8-47d5-8fe3-20eadf06e0f7" (UID: "96b62643-30a8-47d5-8fe3-20eadf06e0f7"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.355102 4696 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/96b62643-30a8-47d5-8fe3-20eadf06e0f7-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.355137 4696 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/96b62643-30a8-47d5-8fe3-20eadf06e0f7-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.355147 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj4n5\" (UniqueName: \"kubernetes.io/projected/96b62643-30a8-47d5-8fe3-20eadf06e0f7-kube-api-access-pj4n5\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.355158 4696 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96b62643-30a8-47d5-8fe3-20eadf06e0f7-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.355167 4696 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/96b62643-30a8-47d5-8fe3-20eadf06e0f7-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.355175 4696 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/96b62643-30a8-47d5-8fe3-20eadf06e0f7-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.355183 4696 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/96b62643-30a8-47d5-8fe3-20eadf06e0f7-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.708834 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-c46nr" Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.741515 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c46nr"] Mar 21 08:33:59 crc kubenswrapper[4696]: I0321 08:33:59.748493 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c46nr"] Mar 21 08:34:00 crc kubenswrapper[4696]: I0321 08:34:00.127438 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568034-hmpdb"] Mar 21 08:34:00 crc kubenswrapper[4696]: E0321 08:34:00.127894 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96b62643-30a8-47d5-8fe3-20eadf06e0f7" containerName="registry" Mar 21 08:34:00 crc kubenswrapper[4696]: I0321 08:34:00.127963 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="96b62643-30a8-47d5-8fe3-20eadf06e0f7" containerName="registry" Mar 21 08:34:00 crc kubenswrapper[4696]: I0321 08:34:00.128275 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="96b62643-30a8-47d5-8fe3-20eadf06e0f7" containerName="registry" Mar 21 08:34:00 crc kubenswrapper[4696]: I0321 08:34:00.128750 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568034-hmpdb" Mar 21 08:34:00 crc kubenswrapper[4696]: I0321 08:34:00.130760 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 08:34:00 crc kubenswrapper[4696]: I0321 08:34:00.130996 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 08:34:00 crc kubenswrapper[4696]: I0321 08:34:00.131210 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 08:34:00 crc kubenswrapper[4696]: I0321 08:34:00.135887 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568034-hmpdb"] Mar 21 08:34:00 crc kubenswrapper[4696]: I0321 08:34:00.164988 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4jgl\" (UniqueName: \"kubernetes.io/projected/6fda0225-3507-4265-aae2-8e0e2f8c11d9-kube-api-access-p4jgl\") pod \"auto-csr-approver-29568034-hmpdb\" (UID: \"6fda0225-3507-4265-aae2-8e0e2f8c11d9\") " pod="openshift-infra/auto-csr-approver-29568034-hmpdb" Mar 21 08:34:00 crc kubenswrapper[4696]: I0321 08:34:00.266174 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4jgl\" (UniqueName: \"kubernetes.io/projected/6fda0225-3507-4265-aae2-8e0e2f8c11d9-kube-api-access-p4jgl\") pod \"auto-csr-approver-29568034-hmpdb\" (UID: \"6fda0225-3507-4265-aae2-8e0e2f8c11d9\") " pod="openshift-infra/auto-csr-approver-29568034-hmpdb" Mar 21 08:34:00 crc kubenswrapper[4696]: I0321 08:34:00.281747 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4jgl\" (UniqueName: \"kubernetes.io/projected/6fda0225-3507-4265-aae2-8e0e2f8c11d9-kube-api-access-p4jgl\") pod \"auto-csr-approver-29568034-hmpdb\" (UID: \"6fda0225-3507-4265-aae2-8e0e2f8c11d9\") " pod="openshift-infra/auto-csr-approver-29568034-hmpdb" Mar 21 08:34:00 crc kubenswrapper[4696]: I0321 08:34:00.452626 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568034-hmpdb" Mar 21 08:34:00 crc kubenswrapper[4696]: I0321 08:34:00.545537 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b62643-30a8-47d5-8fe3-20eadf06e0f7" path="/var/lib/kubelet/pods/96b62643-30a8-47d5-8fe3-20eadf06e0f7/volumes" Mar 21 08:34:00 crc kubenswrapper[4696]: I0321 08:34:00.656803 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568034-hmpdb"] Mar 21 08:34:00 crc kubenswrapper[4696]: W0321 08:34:00.674949 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6fda0225_3507_4265_aae2_8e0e2f8c11d9.slice/crio-ef793a2795813da8f7bfb9384adcbc4e42490d8b5386ea34ff0ab4804e5f127c WatchSource:0}: Error finding container ef793a2795813da8f7bfb9384adcbc4e42490d8b5386ea34ff0ab4804e5f127c: Status 404 returned error can't find the container with id ef793a2795813da8f7bfb9384adcbc4e42490d8b5386ea34ff0ab4804e5f127c Mar 21 08:34:00 crc kubenswrapper[4696]: I0321 08:34:00.713921 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568034-hmpdb" event={"ID":"6fda0225-3507-4265-aae2-8e0e2f8c11d9","Type":"ContainerStarted","Data":"ef793a2795813da8f7bfb9384adcbc4e42490d8b5386ea34ff0ab4804e5f127c"} Mar 21 08:34:01 crc kubenswrapper[4696]: I0321 08:34:01.719146 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568034-hmpdb" event={"ID":"6fda0225-3507-4265-aae2-8e0e2f8c11d9","Type":"ContainerStarted","Data":"c84e974a0e7a698bfe0737e0d5841ca77b33de4f4faeefc53886d83dc6e5eea7"} Mar 21 08:34:01 crc kubenswrapper[4696]: I0321 08:34:01.733847 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29568034-hmpdb" podStartSLOduration=0.978468764 podStartE2EDuration="1.73380399s" podCreationTimestamp="2026-03-21 08:34:00 +0000 UTC" firstStartedPulling="2026-03-21 08:34:00.676442068 +0000 UTC m=+374.797322781" lastFinishedPulling="2026-03-21 08:34:01.431777294 +0000 UTC m=+375.552658007" observedRunningTime="2026-03-21 08:34:01.73069525 +0000 UTC m=+375.851575983" watchObservedRunningTime="2026-03-21 08:34:01.73380399 +0000 UTC m=+375.854684703" Mar 21 08:34:02 crc kubenswrapper[4696]: I0321 08:34:02.726491 4696 generic.go:334] "Generic (PLEG): container finished" podID="6fda0225-3507-4265-aae2-8e0e2f8c11d9" containerID="c84e974a0e7a698bfe0737e0d5841ca77b33de4f4faeefc53886d83dc6e5eea7" exitCode=0 Mar 21 08:34:02 crc kubenswrapper[4696]: I0321 08:34:02.726552 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568034-hmpdb" event={"ID":"6fda0225-3507-4265-aae2-8e0e2f8c11d9","Type":"ContainerDied","Data":"c84e974a0e7a698bfe0737e0d5841ca77b33de4f4faeefc53886d83dc6e5eea7"} Mar 21 08:34:04 crc kubenswrapper[4696]: I0321 08:34:04.029286 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568034-hmpdb" Mar 21 08:34:04 crc kubenswrapper[4696]: I0321 08:34:04.220598 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4jgl\" (UniqueName: \"kubernetes.io/projected/6fda0225-3507-4265-aae2-8e0e2f8c11d9-kube-api-access-p4jgl\") pod \"6fda0225-3507-4265-aae2-8e0e2f8c11d9\" (UID: \"6fda0225-3507-4265-aae2-8e0e2f8c11d9\") " Mar 21 08:34:04 crc kubenswrapper[4696]: I0321 08:34:04.225861 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fda0225-3507-4265-aae2-8e0e2f8c11d9-kube-api-access-p4jgl" (OuterVolumeSpecName: "kube-api-access-p4jgl") pod "6fda0225-3507-4265-aae2-8e0e2f8c11d9" (UID: "6fda0225-3507-4265-aae2-8e0e2f8c11d9"). InnerVolumeSpecName "kube-api-access-p4jgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:34:04 crc kubenswrapper[4696]: I0321 08:34:04.322555 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4jgl\" (UniqueName: \"kubernetes.io/projected/6fda0225-3507-4265-aae2-8e0e2f8c11d9-kube-api-access-p4jgl\") on node \"crc\" DevicePath \"\"" Mar 21 08:34:04 crc kubenswrapper[4696]: I0321 08:34:04.737854 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568034-hmpdb" event={"ID":"6fda0225-3507-4265-aae2-8e0e2f8c11d9","Type":"ContainerDied","Data":"ef793a2795813da8f7bfb9384adcbc4e42490d8b5386ea34ff0ab4804e5f127c"} Mar 21 08:34:04 crc kubenswrapper[4696]: I0321 08:34:04.737895 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef793a2795813da8f7bfb9384adcbc4e42490d8b5386ea34ff0ab4804e5f127c" Mar 21 08:34:04 crc kubenswrapper[4696]: I0321 08:34:04.737932 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568034-hmpdb" Mar 21 08:34:30 crc kubenswrapper[4696]: I0321 08:34:30.340936 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:34:30 crc kubenswrapper[4696]: I0321 08:34:30.341425 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:35:00 crc kubenswrapper[4696]: I0321 08:35:00.341121 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:35:00 crc kubenswrapper[4696]: I0321 08:35:00.341583 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:35:30 crc kubenswrapper[4696]: I0321 08:35:30.342036 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:35:30 crc kubenswrapper[4696]: I0321 08:35:30.342636 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:35:30 crc kubenswrapper[4696]: I0321 08:35:30.342698 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:35:30 crc kubenswrapper[4696]: I0321 08:35:30.343568 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"89c4ac42755bc23804bb46eebf22b6cb1af3d359889a73dd951e91867e71e087"} pod="openshift-machine-config-operator/machine-config-daemon-z7srw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 08:35:30 crc kubenswrapper[4696]: I0321 08:35:30.343669 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" containerID="cri-o://89c4ac42755bc23804bb46eebf22b6cb1af3d359889a73dd951e91867e71e087" gracePeriod=600 Mar 21 08:35:31 crc kubenswrapper[4696]: I0321 08:35:31.213630 4696 generic.go:334] "Generic (PLEG): container finished" podID="daaf227a-2305-495e-8495-a280abcd8e10" containerID="89c4ac42755bc23804bb46eebf22b6cb1af3d359889a73dd951e91867e71e087" exitCode=0 Mar 21 08:35:31 crc kubenswrapper[4696]: I0321 08:35:31.213746 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerDied","Data":"89c4ac42755bc23804bb46eebf22b6cb1af3d359889a73dd951e91867e71e087"} Mar 21 08:35:31 crc kubenswrapper[4696]: I0321 08:35:31.214257 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerStarted","Data":"fc18c7d093fe04a689c494e8d2c8a4d315eb2326b2edffdf3f61d7449dee30d2"} Mar 21 08:35:31 crc kubenswrapper[4696]: I0321 08:35:31.214282 4696 scope.go:117] "RemoveContainer" containerID="9606dfcaf248d5014358578adf21a000d5fb5d698e3980a6ae0ec673dfb6ef85" Mar 21 08:36:00 crc kubenswrapper[4696]: I0321 08:36:00.138407 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568036-lv7tq"] Mar 21 08:36:00 crc kubenswrapper[4696]: E0321 08:36:00.139666 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fda0225-3507-4265-aae2-8e0e2f8c11d9" containerName="oc" Mar 21 08:36:00 crc kubenswrapper[4696]: I0321 08:36:00.139684 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fda0225-3507-4265-aae2-8e0e2f8c11d9" containerName="oc" Mar 21 08:36:00 crc kubenswrapper[4696]: I0321 08:36:00.139846 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fda0225-3507-4265-aae2-8e0e2f8c11d9" containerName="oc" Mar 21 08:36:00 crc kubenswrapper[4696]: I0321 08:36:00.140313 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568036-lv7tq" Mar 21 08:36:00 crc kubenswrapper[4696]: I0321 08:36:00.142981 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 08:36:00 crc kubenswrapper[4696]: I0321 08:36:00.142978 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 08:36:00 crc kubenswrapper[4696]: I0321 08:36:00.144001 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 08:36:00 crc kubenswrapper[4696]: I0321 08:36:00.149713 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568036-lv7tq"] Mar 21 08:36:00 crc kubenswrapper[4696]: I0321 08:36:00.208867 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfgd4\" (UniqueName: \"kubernetes.io/projected/d0cac1ea-786f-41a3-8a4c-0d8d02b40b86-kube-api-access-mfgd4\") pod \"auto-csr-approver-29568036-lv7tq\" (UID: \"d0cac1ea-786f-41a3-8a4c-0d8d02b40b86\") " pod="openshift-infra/auto-csr-approver-29568036-lv7tq" Mar 21 08:36:00 crc kubenswrapper[4696]: I0321 08:36:00.310462 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfgd4\" (UniqueName: \"kubernetes.io/projected/d0cac1ea-786f-41a3-8a4c-0d8d02b40b86-kube-api-access-mfgd4\") pod \"auto-csr-approver-29568036-lv7tq\" (UID: \"d0cac1ea-786f-41a3-8a4c-0d8d02b40b86\") " pod="openshift-infra/auto-csr-approver-29568036-lv7tq" Mar 21 08:36:00 crc kubenswrapper[4696]: I0321 08:36:00.341958 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfgd4\" (UniqueName: \"kubernetes.io/projected/d0cac1ea-786f-41a3-8a4c-0d8d02b40b86-kube-api-access-mfgd4\") pod \"auto-csr-approver-29568036-lv7tq\" (UID: \"d0cac1ea-786f-41a3-8a4c-0d8d02b40b86\") " pod="openshift-infra/auto-csr-approver-29568036-lv7tq" Mar 21 08:36:00 crc kubenswrapper[4696]: I0321 08:36:00.462304 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568036-lv7tq" Mar 21 08:36:00 crc kubenswrapper[4696]: I0321 08:36:00.879095 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568036-lv7tq"] Mar 21 08:36:00 crc kubenswrapper[4696]: I0321 08:36:00.897446 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 08:36:01 crc kubenswrapper[4696]: I0321 08:36:01.393543 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568036-lv7tq" event={"ID":"d0cac1ea-786f-41a3-8a4c-0d8d02b40b86","Type":"ContainerStarted","Data":"c376f24c18682eeb94172a5b015cec78cd80b108944635e38af907e26aa4fecd"} Mar 21 08:36:02 crc kubenswrapper[4696]: I0321 08:36:02.399983 4696 generic.go:334] "Generic (PLEG): container finished" podID="d0cac1ea-786f-41a3-8a4c-0d8d02b40b86" containerID="b2fb2bf44f23b85fb9dce261b53e63b215dcba4f7829a2354c3221ed6d402cc3" exitCode=0 Mar 21 08:36:02 crc kubenswrapper[4696]: I0321 08:36:02.400029 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568036-lv7tq" event={"ID":"d0cac1ea-786f-41a3-8a4c-0d8d02b40b86","Type":"ContainerDied","Data":"b2fb2bf44f23b85fb9dce261b53e63b215dcba4f7829a2354c3221ed6d402cc3"} Mar 21 08:36:03 crc kubenswrapper[4696]: I0321 08:36:03.706533 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568036-lv7tq" Mar 21 08:36:03 crc kubenswrapper[4696]: I0321 08:36:03.783804 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfgd4\" (UniqueName: \"kubernetes.io/projected/d0cac1ea-786f-41a3-8a4c-0d8d02b40b86-kube-api-access-mfgd4\") pod \"d0cac1ea-786f-41a3-8a4c-0d8d02b40b86\" (UID: \"d0cac1ea-786f-41a3-8a4c-0d8d02b40b86\") " Mar 21 08:36:03 crc kubenswrapper[4696]: I0321 08:36:03.793067 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0cac1ea-786f-41a3-8a4c-0d8d02b40b86-kube-api-access-mfgd4" (OuterVolumeSpecName: "kube-api-access-mfgd4") pod "d0cac1ea-786f-41a3-8a4c-0d8d02b40b86" (UID: "d0cac1ea-786f-41a3-8a4c-0d8d02b40b86"). InnerVolumeSpecName "kube-api-access-mfgd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:36:03 crc kubenswrapper[4696]: I0321 08:36:03.885586 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfgd4\" (UniqueName: \"kubernetes.io/projected/d0cac1ea-786f-41a3-8a4c-0d8d02b40b86-kube-api-access-mfgd4\") on node \"crc\" DevicePath \"\"" Mar 21 08:36:04 crc kubenswrapper[4696]: I0321 08:36:04.412593 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568036-lv7tq" event={"ID":"d0cac1ea-786f-41a3-8a4c-0d8d02b40b86","Type":"ContainerDied","Data":"c376f24c18682eeb94172a5b015cec78cd80b108944635e38af907e26aa4fecd"} Mar 21 08:36:04 crc kubenswrapper[4696]: I0321 08:36:04.412656 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c376f24c18682eeb94172a5b015cec78cd80b108944635e38af907e26aa4fecd" Mar 21 08:36:04 crc kubenswrapper[4696]: I0321 08:36:04.412604 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568036-lv7tq" Mar 21 08:36:04 crc kubenswrapper[4696]: I0321 08:36:04.764798 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568030-4x55c"] Mar 21 08:36:04 crc kubenswrapper[4696]: I0321 08:36:04.768240 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568030-4x55c"] Mar 21 08:36:06 crc kubenswrapper[4696]: I0321 08:36:06.540983 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6581056-7677-491f-8f34-5626b43b6e77" path="/var/lib/kubelet/pods/a6581056-7677-491f-8f34-5626b43b6e77/volumes" Mar 21 08:36:46 crc kubenswrapper[4696]: I0321 08:36:46.866419 4696 scope.go:117] "RemoveContainer" containerID="604fb7efc7658349c2c84a277fbbb4ba36f5ca4ba8e34e60ccebde2f771fc479" Mar 21 08:37:30 crc kubenswrapper[4696]: I0321 08:37:30.341748 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:37:30 crc kubenswrapper[4696]: I0321 08:37:30.342286 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:37:46 crc kubenswrapper[4696]: I0321 08:37:46.910765 4696 scope.go:117] "RemoveContainer" containerID="89cca28a71bf3998ea6e1b5417a80f0fb41aadfedf5f4beb2237ac7e4ee7ae9e" Mar 21 08:38:00 crc kubenswrapper[4696]: I0321 08:38:00.154143 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568038-gss7n"] Mar 21 08:38:00 crc kubenswrapper[4696]: E0321 08:38:00.155255 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0cac1ea-786f-41a3-8a4c-0d8d02b40b86" containerName="oc" Mar 21 08:38:00 crc kubenswrapper[4696]: I0321 08:38:00.155283 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0cac1ea-786f-41a3-8a4c-0d8d02b40b86" containerName="oc" Mar 21 08:38:00 crc kubenswrapper[4696]: I0321 08:38:00.155539 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0cac1ea-786f-41a3-8a4c-0d8d02b40b86" containerName="oc" Mar 21 08:38:00 crc kubenswrapper[4696]: I0321 08:38:00.157742 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568038-gss7n" Mar 21 08:38:00 crc kubenswrapper[4696]: I0321 08:38:00.160196 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 08:38:00 crc kubenswrapper[4696]: I0321 08:38:00.161140 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 08:38:00 crc kubenswrapper[4696]: I0321 08:38:00.161352 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 08:38:00 crc kubenswrapper[4696]: I0321 08:38:00.177522 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568038-gss7n"] Mar 21 08:38:00 crc kubenswrapper[4696]: I0321 08:38:00.185526 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvl2l\" (UniqueName: \"kubernetes.io/projected/9a11bb77-b5ab-41fb-aea9-5d7de836af84-kube-api-access-fvl2l\") pod \"auto-csr-approver-29568038-gss7n\" (UID: \"9a11bb77-b5ab-41fb-aea9-5d7de836af84\") " pod="openshift-infra/auto-csr-approver-29568038-gss7n" Mar 21 08:38:00 crc kubenswrapper[4696]: I0321 08:38:00.286878 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvl2l\" (UniqueName: \"kubernetes.io/projected/9a11bb77-b5ab-41fb-aea9-5d7de836af84-kube-api-access-fvl2l\") pod \"auto-csr-approver-29568038-gss7n\" (UID: \"9a11bb77-b5ab-41fb-aea9-5d7de836af84\") " pod="openshift-infra/auto-csr-approver-29568038-gss7n" Mar 21 08:38:00 crc kubenswrapper[4696]: I0321 08:38:00.320784 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvl2l\" (UniqueName: \"kubernetes.io/projected/9a11bb77-b5ab-41fb-aea9-5d7de836af84-kube-api-access-fvl2l\") pod \"auto-csr-approver-29568038-gss7n\" (UID: \"9a11bb77-b5ab-41fb-aea9-5d7de836af84\") " pod="openshift-infra/auto-csr-approver-29568038-gss7n" Mar 21 08:38:00 crc kubenswrapper[4696]: I0321 08:38:00.341748 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:38:00 crc kubenswrapper[4696]: I0321 08:38:00.341859 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:38:00 crc kubenswrapper[4696]: I0321 08:38:00.496901 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568038-gss7n" Mar 21 08:38:00 crc kubenswrapper[4696]: I0321 08:38:00.758122 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568038-gss7n"] Mar 21 08:38:01 crc kubenswrapper[4696]: I0321 08:38:01.085976 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568038-gss7n" event={"ID":"9a11bb77-b5ab-41fb-aea9-5d7de836af84","Type":"ContainerStarted","Data":"915b91ecf434f1b52d1d47e6d6a832317168f0b0616c2466bf45bba85c111692"} Mar 21 08:38:02 crc kubenswrapper[4696]: I0321 08:38:02.092027 4696 generic.go:334] "Generic (PLEG): container finished" podID="9a11bb77-b5ab-41fb-aea9-5d7de836af84" containerID="2e1df926844a18d70f9120e60df866892a5d9837c576b458c331c1a881722233" exitCode=0 Mar 21 08:38:02 crc kubenswrapper[4696]: I0321 08:38:02.092270 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568038-gss7n" event={"ID":"9a11bb77-b5ab-41fb-aea9-5d7de836af84","Type":"ContainerDied","Data":"2e1df926844a18d70f9120e60df866892a5d9837c576b458c331c1a881722233"} Mar 21 08:38:03 crc kubenswrapper[4696]: I0321 08:38:03.364086 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568038-gss7n" Mar 21 08:38:03 crc kubenswrapper[4696]: I0321 08:38:03.426679 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvl2l\" (UniqueName: \"kubernetes.io/projected/9a11bb77-b5ab-41fb-aea9-5d7de836af84-kube-api-access-fvl2l\") pod \"9a11bb77-b5ab-41fb-aea9-5d7de836af84\" (UID: \"9a11bb77-b5ab-41fb-aea9-5d7de836af84\") " Mar 21 08:38:03 crc kubenswrapper[4696]: I0321 08:38:03.431700 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a11bb77-b5ab-41fb-aea9-5d7de836af84-kube-api-access-fvl2l" (OuterVolumeSpecName: "kube-api-access-fvl2l") pod "9a11bb77-b5ab-41fb-aea9-5d7de836af84" (UID: "9a11bb77-b5ab-41fb-aea9-5d7de836af84"). InnerVolumeSpecName "kube-api-access-fvl2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:38:03 crc kubenswrapper[4696]: I0321 08:38:03.528195 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvl2l\" (UniqueName: \"kubernetes.io/projected/9a11bb77-b5ab-41fb-aea9-5d7de836af84-kube-api-access-fvl2l\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:04 crc kubenswrapper[4696]: I0321 08:38:04.105629 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568038-gss7n" event={"ID":"9a11bb77-b5ab-41fb-aea9-5d7de836af84","Type":"ContainerDied","Data":"915b91ecf434f1b52d1d47e6d6a832317168f0b0616c2466bf45bba85c111692"} Mar 21 08:38:04 crc kubenswrapper[4696]: I0321 08:38:04.105691 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="915b91ecf434f1b52d1d47e6d6a832317168f0b0616c2466bf45bba85c111692" Mar 21 08:38:04 crc kubenswrapper[4696]: I0321 08:38:04.105772 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568038-gss7n" Mar 21 08:38:04 crc kubenswrapper[4696]: I0321 08:38:04.426149 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568032-ljb7d"] Mar 21 08:38:04 crc kubenswrapper[4696]: I0321 08:38:04.431103 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568032-ljb7d"] Mar 21 08:38:04 crc kubenswrapper[4696]: I0321 08:38:04.545423 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b435a39c-eb1b-4ae9-9cc3-7dc51f632111" path="/var/lib/kubelet/pods/b435a39c-eb1b-4ae9-9cc3-7dc51f632111/volumes" Mar 21 08:38:30 crc kubenswrapper[4696]: I0321 08:38:30.342542 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:38:30 crc kubenswrapper[4696]: I0321 08:38:30.343039 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:38:30 crc kubenswrapper[4696]: I0321 08:38:30.343085 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:38:30 crc kubenswrapper[4696]: I0321 08:38:30.343608 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fc18c7d093fe04a689c494e8d2c8a4d315eb2326b2edffdf3f61d7449dee30d2"} pod="openshift-machine-config-operator/machine-config-daemon-z7srw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 08:38:30 crc kubenswrapper[4696]: I0321 08:38:30.343671 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" containerID="cri-o://fc18c7d093fe04a689c494e8d2c8a4d315eb2326b2edffdf3f61d7449dee30d2" gracePeriod=600 Mar 21 08:38:31 crc kubenswrapper[4696]: I0321 08:38:31.268466 4696 generic.go:334] "Generic (PLEG): container finished" podID="daaf227a-2305-495e-8495-a280abcd8e10" containerID="fc18c7d093fe04a689c494e8d2c8a4d315eb2326b2edffdf3f61d7449dee30d2" exitCode=0 Mar 21 08:38:31 crc kubenswrapper[4696]: I0321 08:38:31.268536 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerDied","Data":"fc18c7d093fe04a689c494e8d2c8a4d315eb2326b2edffdf3f61d7449dee30d2"} Mar 21 08:38:31 crc kubenswrapper[4696]: I0321 08:38:31.269044 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerStarted","Data":"a451c8e7ce2e953d0eead4e37b6cb08fe2df64dae3d013237676d9b8ff9dbb40"} Mar 21 08:38:31 crc kubenswrapper[4696]: I0321 08:38:31.269066 4696 scope.go:117] "RemoveContainer" containerID="89c4ac42755bc23804bb46eebf22b6cb1af3d359889a73dd951e91867e71e087" Mar 21 08:38:33 crc kubenswrapper[4696]: I0321 08:38:33.029920 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5"] Mar 21 08:38:33 crc kubenswrapper[4696]: E0321 08:38:33.030373 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a11bb77-b5ab-41fb-aea9-5d7de836af84" containerName="oc" Mar 21 08:38:33 crc kubenswrapper[4696]: I0321 08:38:33.030384 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a11bb77-b5ab-41fb-aea9-5d7de836af84" containerName="oc" Mar 21 08:38:33 crc kubenswrapper[4696]: I0321 08:38:33.030477 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a11bb77-b5ab-41fb-aea9-5d7de836af84" containerName="oc" Mar 21 08:38:33 crc kubenswrapper[4696]: I0321 08:38:33.031182 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5" Mar 21 08:38:33 crc kubenswrapper[4696]: I0321 08:38:33.033523 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 21 08:38:33 crc kubenswrapper[4696]: I0321 08:38:33.044799 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5"] Mar 21 08:38:33 crc kubenswrapper[4696]: I0321 08:38:33.094457 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e4b15b92-b219-4ad6-a9ab-7d05a7130c92-bundle\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5\" (UID: \"e4b15b92-b219-4ad6-a9ab-7d05a7130c92\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5" Mar 21 08:38:33 crc kubenswrapper[4696]: I0321 08:38:33.094528 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q487c\" (UniqueName: \"kubernetes.io/projected/e4b15b92-b219-4ad6-a9ab-7d05a7130c92-kube-api-access-q487c\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5\" (UID: \"e4b15b92-b219-4ad6-a9ab-7d05a7130c92\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5" Mar 21 08:38:33 crc kubenswrapper[4696]: I0321 08:38:33.094592 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e4b15b92-b219-4ad6-a9ab-7d05a7130c92-util\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5\" (UID: \"e4b15b92-b219-4ad6-a9ab-7d05a7130c92\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5" Mar 21 08:38:33 crc kubenswrapper[4696]: I0321 08:38:33.196056 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e4b15b92-b219-4ad6-a9ab-7d05a7130c92-bundle\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5\" (UID: \"e4b15b92-b219-4ad6-a9ab-7d05a7130c92\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5" Mar 21 08:38:33 crc kubenswrapper[4696]: I0321 08:38:33.196131 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q487c\" (UniqueName: \"kubernetes.io/projected/e4b15b92-b219-4ad6-a9ab-7d05a7130c92-kube-api-access-q487c\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5\" (UID: \"e4b15b92-b219-4ad6-a9ab-7d05a7130c92\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5" Mar 21 08:38:33 crc kubenswrapper[4696]: I0321 08:38:33.196177 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e4b15b92-b219-4ad6-a9ab-7d05a7130c92-util\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5\" (UID: \"e4b15b92-b219-4ad6-a9ab-7d05a7130c92\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5" Mar 21 08:38:33 crc kubenswrapper[4696]: I0321 08:38:33.196735 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e4b15b92-b219-4ad6-a9ab-7d05a7130c92-util\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5\" (UID: \"e4b15b92-b219-4ad6-a9ab-7d05a7130c92\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5" Mar 21 08:38:33 crc kubenswrapper[4696]: I0321 08:38:33.196851 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e4b15b92-b219-4ad6-a9ab-7d05a7130c92-bundle\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5\" (UID: \"e4b15b92-b219-4ad6-a9ab-7d05a7130c92\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5" Mar 21 08:38:33 crc kubenswrapper[4696]: I0321 08:38:33.216083 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q487c\" (UniqueName: \"kubernetes.io/projected/e4b15b92-b219-4ad6-a9ab-7d05a7130c92-kube-api-access-q487c\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5\" (UID: \"e4b15b92-b219-4ad6-a9ab-7d05a7130c92\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5" Mar 21 08:38:33 crc kubenswrapper[4696]: I0321 08:38:33.345408 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5" Mar 21 08:38:33 crc kubenswrapper[4696]: I0321 08:38:33.533608 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5"] Mar 21 08:38:34 crc kubenswrapper[4696]: I0321 08:38:34.290458 4696 generic.go:334] "Generic (PLEG): container finished" podID="e4b15b92-b219-4ad6-a9ab-7d05a7130c92" containerID="f43e07c281f20b2057cdf1fb0307a04661c3f869add0152b73028917c3a383d7" exitCode=0 Mar 21 08:38:34 crc kubenswrapper[4696]: I0321 08:38:34.290571 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5" event={"ID":"e4b15b92-b219-4ad6-a9ab-7d05a7130c92","Type":"ContainerDied","Data":"f43e07c281f20b2057cdf1fb0307a04661c3f869add0152b73028917c3a383d7"} Mar 21 08:38:34 crc kubenswrapper[4696]: I0321 08:38:34.290855 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5" event={"ID":"e4b15b92-b219-4ad6-a9ab-7d05a7130c92","Type":"ContainerStarted","Data":"c53652b427104eaefbfc326826743b01d58e1e3acd8c38e638e027ef2fabab4a"} Mar 21 08:38:36 crc kubenswrapper[4696]: I0321 08:38:36.306116 4696 generic.go:334] "Generic (PLEG): container finished" podID="e4b15b92-b219-4ad6-a9ab-7d05a7130c92" containerID="ba40cb6e988a2c46fa8b15ad87d52ec7f7f39111e76b64804f82ec94172fa1e4" exitCode=0 Mar 21 08:38:36 crc kubenswrapper[4696]: I0321 08:38:36.306171 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5" event={"ID":"e4b15b92-b219-4ad6-a9ab-7d05a7130c92","Type":"ContainerDied","Data":"ba40cb6e988a2c46fa8b15ad87d52ec7f7f39111e76b64804f82ec94172fa1e4"} Mar 21 08:38:37 crc kubenswrapper[4696]: I0321 08:38:37.316062 4696 generic.go:334] "Generic (PLEG): container finished" podID="e4b15b92-b219-4ad6-a9ab-7d05a7130c92" containerID="16ff84258d6895ceecd27be807aa9f5a6dc76462273b139988f2b6f1852cccea" exitCode=0 Mar 21 08:38:37 crc kubenswrapper[4696]: I0321 08:38:37.316100 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5" event={"ID":"e4b15b92-b219-4ad6-a9ab-7d05a7130c92","Type":"ContainerDied","Data":"16ff84258d6895ceecd27be807aa9f5a6dc76462273b139988f2b6f1852cccea"} Mar 21 08:38:38 crc kubenswrapper[4696]: I0321 08:38:38.570861 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5" Mar 21 08:38:38 crc kubenswrapper[4696]: I0321 08:38:38.759208 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q487c\" (UniqueName: \"kubernetes.io/projected/e4b15b92-b219-4ad6-a9ab-7d05a7130c92-kube-api-access-q487c\") pod \"e4b15b92-b219-4ad6-a9ab-7d05a7130c92\" (UID: \"e4b15b92-b219-4ad6-a9ab-7d05a7130c92\") " Mar 21 08:38:38 crc kubenswrapper[4696]: I0321 08:38:38.759335 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e4b15b92-b219-4ad6-a9ab-7d05a7130c92-bundle\") pod \"e4b15b92-b219-4ad6-a9ab-7d05a7130c92\" (UID: \"e4b15b92-b219-4ad6-a9ab-7d05a7130c92\") " Mar 21 08:38:38 crc kubenswrapper[4696]: I0321 08:38:38.759373 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e4b15b92-b219-4ad6-a9ab-7d05a7130c92-util\") pod \"e4b15b92-b219-4ad6-a9ab-7d05a7130c92\" (UID: \"e4b15b92-b219-4ad6-a9ab-7d05a7130c92\") " Mar 21 08:38:38 crc kubenswrapper[4696]: I0321 08:38:38.762543 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4b15b92-b219-4ad6-a9ab-7d05a7130c92-bundle" (OuterVolumeSpecName: "bundle") pod "e4b15b92-b219-4ad6-a9ab-7d05a7130c92" (UID: "e4b15b92-b219-4ad6-a9ab-7d05a7130c92"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:38:38 crc kubenswrapper[4696]: I0321 08:38:38.764992 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4b15b92-b219-4ad6-a9ab-7d05a7130c92-kube-api-access-q487c" (OuterVolumeSpecName: "kube-api-access-q487c") pod "e4b15b92-b219-4ad6-a9ab-7d05a7130c92" (UID: "e4b15b92-b219-4ad6-a9ab-7d05a7130c92"). InnerVolumeSpecName "kube-api-access-q487c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:38:38 crc kubenswrapper[4696]: I0321 08:38:38.775659 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4b15b92-b219-4ad6-a9ab-7d05a7130c92-util" (OuterVolumeSpecName: "util") pod "e4b15b92-b219-4ad6-a9ab-7d05a7130c92" (UID: "e4b15b92-b219-4ad6-a9ab-7d05a7130c92"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:38:38 crc kubenswrapper[4696]: I0321 08:38:38.861537 4696 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e4b15b92-b219-4ad6-a9ab-7d05a7130c92-util\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:38 crc kubenswrapper[4696]: I0321 08:38:38.861607 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q487c\" (UniqueName: \"kubernetes.io/projected/e4b15b92-b219-4ad6-a9ab-7d05a7130c92-kube-api-access-q487c\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:38 crc kubenswrapper[4696]: I0321 08:38:38.861658 4696 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e4b15b92-b219-4ad6-a9ab-7d05a7130c92-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:39 crc kubenswrapper[4696]: I0321 08:38:39.329381 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5" event={"ID":"e4b15b92-b219-4ad6-a9ab-7d05a7130c92","Type":"ContainerDied","Data":"c53652b427104eaefbfc326826743b01d58e1e3acd8c38e638e027ef2fabab4a"} Mar 21 08:38:39 crc kubenswrapper[4696]: I0321 08:38:39.329415 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c53652b427104eaefbfc326826743b01d58e1e3acd8c38e638e027ef2fabab4a" Mar 21 08:38:39 crc kubenswrapper[4696]: I0321 08:38:39.329422 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5" Mar 21 08:38:44 crc kubenswrapper[4696]: I0321 08:38:44.276637 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9tp76"] Mar 21 08:38:44 crc kubenswrapper[4696]: I0321 08:38:44.277435 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="ovn-controller" containerID="cri-o://42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054" gracePeriod=30 Mar 21 08:38:44 crc kubenswrapper[4696]: I0321 08:38:44.277795 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="sbdb" containerID="cri-o://7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0" gracePeriod=30 Mar 21 08:38:44 crc kubenswrapper[4696]: I0321 08:38:44.277857 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="nbdb" containerID="cri-o://4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8" gracePeriod=30 Mar 21 08:38:44 crc kubenswrapper[4696]: I0321 08:38:44.277885 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="northd" containerID="cri-o://beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a" gracePeriod=30 Mar 21 08:38:44 crc kubenswrapper[4696]: I0321 08:38:44.277914 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a" gracePeriod=30 Mar 21 08:38:44 crc kubenswrapper[4696]: I0321 08:38:44.277939 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="kube-rbac-proxy-node" containerID="cri-o://cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7" gracePeriod=30 Mar 21 08:38:44 crc kubenswrapper[4696]: I0321 08:38:44.277971 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="ovn-acl-logging" containerID="cri-o://92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59" gracePeriod=30 Mar 21 08:38:44 crc kubenswrapper[4696]: I0321 08:38:44.326408 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="ovnkube-controller" containerID="cri-o://4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281" gracePeriod=30 Mar 21 08:38:44 crc kubenswrapper[4696]: I0321 08:38:44.972567 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9tp76_9f0e7707-cb43-4f15-b309-7f651da35ddf/ovn-acl-logging/0.log" Mar 21 08:38:44 crc kubenswrapper[4696]: I0321 08:38:44.973372 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9tp76_9f0e7707-cb43-4f15-b309-7f651da35ddf/ovn-controller/0.log" Mar 21 08:38:44 crc kubenswrapper[4696]: I0321 08:38:44.973790 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.040983 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-klgqj"] Mar 21 08:38:45 crc kubenswrapper[4696]: E0321 08:38:45.041238 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="northd" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.041252 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="northd" Mar 21 08:38:45 crc kubenswrapper[4696]: E0321 08:38:45.041263 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="ovn-controller" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.041268 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="ovn-controller" Mar 21 08:38:45 crc kubenswrapper[4696]: E0321 08:38:45.041275 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4b15b92-b219-4ad6-a9ab-7d05a7130c92" containerName="util" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.041281 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4b15b92-b219-4ad6-a9ab-7d05a7130c92" containerName="util" Mar 21 08:38:45 crc kubenswrapper[4696]: E0321 08:38:45.041292 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="kubecfg-setup" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.041297 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="kubecfg-setup" Mar 21 08:38:45 crc kubenswrapper[4696]: E0321 08:38:45.041309 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4b15b92-b219-4ad6-a9ab-7d05a7130c92" containerName="pull" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.041314 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4b15b92-b219-4ad6-a9ab-7d05a7130c92" containerName="pull" Mar 21 08:38:45 crc kubenswrapper[4696]: E0321 08:38:45.041320 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="ovn-acl-logging" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.041326 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="ovn-acl-logging" Mar 21 08:38:45 crc kubenswrapper[4696]: E0321 08:38:45.041333 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="kube-rbac-proxy-ovn-metrics" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.041338 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="kube-rbac-proxy-ovn-metrics" Mar 21 08:38:45 crc kubenswrapper[4696]: E0321 08:38:45.041347 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="ovnkube-controller" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.041353 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="ovnkube-controller" Mar 21 08:38:45 crc kubenswrapper[4696]: E0321 08:38:45.041360 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="sbdb" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.041366 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="sbdb" Mar 21 08:38:45 crc kubenswrapper[4696]: E0321 08:38:45.041374 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="kube-rbac-proxy-node" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.041379 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="kube-rbac-proxy-node" Mar 21 08:38:45 crc kubenswrapper[4696]: E0321 08:38:45.041388 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4b15b92-b219-4ad6-a9ab-7d05a7130c92" containerName="extract" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.041394 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4b15b92-b219-4ad6-a9ab-7d05a7130c92" containerName="extract" Mar 21 08:38:45 crc kubenswrapper[4696]: E0321 08:38:45.041400 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="nbdb" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.041407 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="nbdb" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.041508 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="ovn-controller" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.041518 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="sbdb" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.041530 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="kube-rbac-proxy-node" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.041537 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="northd" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.041542 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="nbdb" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.041548 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="ovnkube-controller" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.041555 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="kube-rbac-proxy-ovn-metrics" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.041565 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerName="ovn-acl-logging" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.041573 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4b15b92-b219-4ad6-a9ab-7d05a7130c92" containerName="extract" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.043290 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.143637 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-etc-openvswitch\") pod \"9f0e7707-cb43-4f15-b309-7f651da35ddf\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.143929 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-cni-bin\") pod \"9f0e7707-cb43-4f15-b309-7f651da35ddf\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144068 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-var-lib-cni-networks-ovn-kubernetes\") pod \"9f0e7707-cb43-4f15-b309-7f651da35ddf\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.143770 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "9f0e7707-cb43-4f15-b309-7f651da35ddf" (UID: "9f0e7707-cb43-4f15-b309-7f651da35ddf"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.143999 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "9f0e7707-cb43-4f15-b309-7f651da35ddf" (UID: "9f0e7707-cb43-4f15-b309-7f651da35ddf"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144128 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "9f0e7707-cb43-4f15-b309-7f651da35ddf" (UID: "9f0e7707-cb43-4f15-b309-7f651da35ddf"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144263 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "9f0e7707-cb43-4f15-b309-7f651da35ddf" (UID: "9f0e7707-cb43-4f15-b309-7f651da35ddf"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144151 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-run-ovn-kubernetes\") pod \"9f0e7707-cb43-4f15-b309-7f651da35ddf\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144308 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-cni-netd\") pod \"9f0e7707-cb43-4f15-b309-7f651da35ddf\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144335 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9f0e7707-cb43-4f15-b309-7f651da35ddf-ovnkube-script-lib\") pod \"9f0e7707-cb43-4f15-b309-7f651da35ddf\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144369 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-log-socket\") pod \"9f0e7707-cb43-4f15-b309-7f651da35ddf\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144396 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-systemd-units\") pod \"9f0e7707-cb43-4f15-b309-7f651da35ddf\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144416 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-run-netns\") pod \"9f0e7707-cb43-4f15-b309-7f651da35ddf\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144445 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-run-systemd\") pod \"9f0e7707-cb43-4f15-b309-7f651da35ddf\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144480 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-run-openvswitch\") pod \"9f0e7707-cb43-4f15-b309-7f651da35ddf\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144506 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stzpr\" (UniqueName: \"kubernetes.io/projected/9f0e7707-cb43-4f15-b309-7f651da35ddf-kube-api-access-stzpr\") pod \"9f0e7707-cb43-4f15-b309-7f651da35ddf\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144582 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-slash\") pod \"9f0e7707-cb43-4f15-b309-7f651da35ddf\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144607 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9f0e7707-cb43-4f15-b309-7f651da35ddf-ovn-node-metrics-cert\") pod \"9f0e7707-cb43-4f15-b309-7f651da35ddf\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144652 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9f0e7707-cb43-4f15-b309-7f651da35ddf-env-overrides\") pod \"9f0e7707-cb43-4f15-b309-7f651da35ddf\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144675 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-var-lib-openvswitch\") pod \"9f0e7707-cb43-4f15-b309-7f651da35ddf\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144711 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9f0e7707-cb43-4f15-b309-7f651da35ddf-ovnkube-config\") pod \"9f0e7707-cb43-4f15-b309-7f651da35ddf\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144736 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-kubelet\") pod \"9f0e7707-cb43-4f15-b309-7f651da35ddf\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144763 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-node-log\") pod \"9f0e7707-cb43-4f15-b309-7f651da35ddf\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144783 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-run-ovn\") pod \"9f0e7707-cb43-4f15-b309-7f651da35ddf\" (UID: \"9f0e7707-cb43-4f15-b309-7f651da35ddf\") " Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144970 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-host-slash\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.144986 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "9f0e7707-cb43-4f15-b309-7f651da35ddf" (UID: "9f0e7707-cb43-4f15-b309-7f651da35ddf"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145002 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-log-socket\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145037 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-slash" (OuterVolumeSpecName: "host-slash") pod "9f0e7707-cb43-4f15-b309-7f651da35ddf" (UID: "9f0e7707-cb43-4f15-b309-7f651da35ddf"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145063 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-etc-openvswitch\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145091 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-host-cni-bin\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145113 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d53ef7be-e64a-4164-b951-03979405c91f-ovnkube-script-lib\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145132 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-host-run-ovn-kubernetes\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145156 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-host-cni-netd\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145172 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-host-run-netns\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145188 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-systemd-units\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145222 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d53ef7be-e64a-4164-b951-03979405c91f-ovn-node-metrics-cert\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145242 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtr2j\" (UniqueName: \"kubernetes.io/projected/d53ef7be-e64a-4164-b951-03979405c91f-kube-api-access-vtr2j\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145257 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d53ef7be-e64a-4164-b951-03979405c91f-ovnkube-config\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145274 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-host-kubelet\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145293 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-var-lib-openvswitch\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145312 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145335 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-run-openvswitch\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145351 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-node-log\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145369 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-run-systemd\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145382 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d53ef7be-e64a-4164-b951-03979405c91f-env-overrides\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145399 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-run-ovn\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145434 4696 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145445 4696 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145455 4696 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145464 4696 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-slash\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145475 4696 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145484 4696 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145513 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "9f0e7707-cb43-4f15-b309-7f651da35ddf" (UID: "9f0e7707-cb43-4f15-b309-7f651da35ddf"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145676 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "9f0e7707-cb43-4f15-b309-7f651da35ddf" (UID: "9f0e7707-cb43-4f15-b309-7f651da35ddf"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145939 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f0e7707-cb43-4f15-b309-7f651da35ddf-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "9f0e7707-cb43-4f15-b309-7f651da35ddf" (UID: "9f0e7707-cb43-4f15-b309-7f651da35ddf"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.145971 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-log-socket" (OuterVolumeSpecName: "log-socket") pod "9f0e7707-cb43-4f15-b309-7f651da35ddf" (UID: "9f0e7707-cb43-4f15-b309-7f651da35ddf"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.146036 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f0e7707-cb43-4f15-b309-7f651da35ddf-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "9f0e7707-cb43-4f15-b309-7f651da35ddf" (UID: "9f0e7707-cb43-4f15-b309-7f651da35ddf"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.146065 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "9f0e7707-cb43-4f15-b309-7f651da35ddf" (UID: "9f0e7707-cb43-4f15-b309-7f651da35ddf"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.146089 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "9f0e7707-cb43-4f15-b309-7f651da35ddf" (UID: "9f0e7707-cb43-4f15-b309-7f651da35ddf"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.146127 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-node-log" (OuterVolumeSpecName: "node-log") pod "9f0e7707-cb43-4f15-b309-7f651da35ddf" (UID: "9f0e7707-cb43-4f15-b309-7f651da35ddf"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.146157 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "9f0e7707-cb43-4f15-b309-7f651da35ddf" (UID: "9f0e7707-cb43-4f15-b309-7f651da35ddf"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.146176 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "9f0e7707-cb43-4f15-b309-7f651da35ddf" (UID: "9f0e7707-cb43-4f15-b309-7f651da35ddf"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.146262 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f0e7707-cb43-4f15-b309-7f651da35ddf-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "9f0e7707-cb43-4f15-b309-7f651da35ddf" (UID: "9f0e7707-cb43-4f15-b309-7f651da35ddf"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.155326 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f0e7707-cb43-4f15-b309-7f651da35ddf-kube-api-access-stzpr" (OuterVolumeSpecName: "kube-api-access-stzpr") pod "9f0e7707-cb43-4f15-b309-7f651da35ddf" (UID: "9f0e7707-cb43-4f15-b309-7f651da35ddf"). InnerVolumeSpecName "kube-api-access-stzpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.163232 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f0e7707-cb43-4f15-b309-7f651da35ddf-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "9f0e7707-cb43-4f15-b309-7f651da35ddf" (UID: "9f0e7707-cb43-4f15-b309-7f651da35ddf"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.167001 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "9f0e7707-cb43-4f15-b309-7f651da35ddf" (UID: "9f0e7707-cb43-4f15-b309-7f651da35ddf"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.246852 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-etc-openvswitch\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.246902 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-host-cni-bin\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.246927 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d53ef7be-e64a-4164-b951-03979405c91f-ovnkube-script-lib\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.246944 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-host-run-ovn-kubernetes\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.246979 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-host-cni-netd\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247000 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-host-run-netns\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247018 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-systemd-units\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247033 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-host-cni-bin\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247063 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-host-run-ovn-kubernetes\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247046 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-host-cni-netd\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247011 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-etc-openvswitch\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247111 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-systemd-units\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247065 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d53ef7be-e64a-4164-b951-03979405c91f-ovn-node-metrics-cert\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247156 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtr2j\" (UniqueName: \"kubernetes.io/projected/d53ef7be-e64a-4164-b951-03979405c91f-kube-api-access-vtr2j\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247174 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d53ef7be-e64a-4164-b951-03979405c91f-ovnkube-config\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247192 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-host-kubelet\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247209 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-var-lib-openvswitch\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247229 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247251 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-run-openvswitch\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247271 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-node-log\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247288 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-run-systemd\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247303 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d53ef7be-e64a-4164-b951-03979405c91f-env-overrides\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247319 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-run-ovn\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247340 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-log-socket\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247356 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-host-slash\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247405 4696 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247416 4696 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247424 4696 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247433 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stzpr\" (UniqueName: \"kubernetes.io/projected/9f0e7707-cb43-4f15-b309-7f651da35ddf-kube-api-access-stzpr\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247446 4696 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9f0e7707-cb43-4f15-b309-7f651da35ddf-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247455 4696 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9f0e7707-cb43-4f15-b309-7f651da35ddf-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247463 4696 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247472 4696 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9f0e7707-cb43-4f15-b309-7f651da35ddf-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247480 4696 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247487 4696 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-node-log\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247496 4696 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247504 4696 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247512 4696 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9f0e7707-cb43-4f15-b309-7f651da35ddf-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247520 4696 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9f0e7707-cb43-4f15-b309-7f651da35ddf-log-socket\") on node \"crc\" DevicePath \"\"" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247548 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-host-slash\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247134 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-host-run-netns\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247699 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d53ef7be-e64a-4164-b951-03979405c91f-ovnkube-script-lib\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247759 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-run-openvswitch\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247843 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-node-log\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.247871 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-run-systemd\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.248214 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d53ef7be-e64a-4164-b951-03979405c91f-ovnkube-config\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.248252 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-host-kubelet\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.248276 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-var-lib-openvswitch\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.248300 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.248327 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-run-ovn\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.248333 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d53ef7be-e64a-4164-b951-03979405c91f-env-overrides\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.248351 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d53ef7be-e64a-4164-b951-03979405c91f-log-socket\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.252480 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d53ef7be-e64a-4164-b951-03979405c91f-ovn-node-metrics-cert\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.264308 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtr2j\" (UniqueName: \"kubernetes.io/projected/d53ef7be-e64a-4164-b951-03979405c91f-kube-api-access-vtr2j\") pod \"ovnkube-node-klgqj\" (UID: \"d53ef7be-e64a-4164-b951-03979405c91f\") " pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.357239 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.368975 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9tp76_9f0e7707-cb43-4f15-b309-7f651da35ddf/ovn-acl-logging/0.log" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.369837 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9tp76_9f0e7707-cb43-4f15-b309-7f651da35ddf/ovn-controller/0.log" Mar 21 08:38:45 crc kubenswrapper[4696]: W0321 08:38:45.375269 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd53ef7be_e64a_4164_b951_03979405c91f.slice/crio-9ee45c48c08498c51221780230e50a28570c2224c2b7cd420719747511e2af3e WatchSource:0}: Error finding container 9ee45c48c08498c51221780230e50a28570c2224c2b7cd420719747511e2af3e: Status 404 returned error can't find the container with id 9ee45c48c08498c51221780230e50a28570c2224c2b7cd420719747511e2af3e Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375459 4696 generic.go:334] "Generic (PLEG): container finished" podID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerID="4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281" exitCode=0 Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375497 4696 generic.go:334] "Generic (PLEG): container finished" podID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerID="7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0" exitCode=0 Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375508 4696 generic.go:334] "Generic (PLEG): container finished" podID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerID="4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8" exitCode=0 Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375517 4696 generic.go:334] "Generic (PLEG): container finished" podID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerID="beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a" exitCode=0 Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375526 4696 generic.go:334] "Generic (PLEG): container finished" podID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerID="2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a" exitCode=0 Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375535 4696 generic.go:334] "Generic (PLEG): container finished" podID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerID="cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7" exitCode=0 Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375544 4696 generic.go:334] "Generic (PLEG): container finished" podID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerID="92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59" exitCode=143 Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375550 4696 generic.go:334] "Generic (PLEG): container finished" podID="9f0e7707-cb43-4f15-b309-7f651da35ddf" containerID="42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054" exitCode=143 Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375554 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375511 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" event={"ID":"9f0e7707-cb43-4f15-b309-7f651da35ddf","Type":"ContainerDied","Data":"4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375634 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" event={"ID":"9f0e7707-cb43-4f15-b309-7f651da35ddf","Type":"ContainerDied","Data":"7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375650 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" event={"ID":"9f0e7707-cb43-4f15-b309-7f651da35ddf","Type":"ContainerDied","Data":"4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375661 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" event={"ID":"9f0e7707-cb43-4f15-b309-7f651da35ddf","Type":"ContainerDied","Data":"beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375670 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" event={"ID":"9f0e7707-cb43-4f15-b309-7f651da35ddf","Type":"ContainerDied","Data":"2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375680 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" event={"ID":"9f0e7707-cb43-4f15-b309-7f651da35ddf","Type":"ContainerDied","Data":"cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375692 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375702 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375708 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375714 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" event={"ID":"9f0e7707-cb43-4f15-b309-7f651da35ddf","Type":"ContainerDied","Data":"92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375715 4696 scope.go:117] "RemoveContainer" containerID="4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375722 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375887 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375902 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375909 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375916 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375922 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375929 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375935 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375941 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375957 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" event={"ID":"9f0e7707-cb43-4f15-b309-7f651da35ddf","Type":"ContainerDied","Data":"42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375979 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375986 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.375993 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.376000 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.376008 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.376016 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.376022 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.376028 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.376036 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.376046 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tp76" event={"ID":"9f0e7707-cb43-4f15-b309-7f651da35ddf","Type":"ContainerDied","Data":"a539a3280d13394f870f8a68e82bffd19b597121e60aeb43f584fe392860c830"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.376057 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.376065 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.376071 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.376078 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.376085 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.376091 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.376097 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.376104 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.376110 4696 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.377532 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-t6d75_42c712ed-b850-4947-b16e-5517c26612f8/kube-multus/0.log" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.377559 4696 generic.go:334] "Generic (PLEG): container finished" podID="42c712ed-b850-4947-b16e-5517c26612f8" containerID="c4d2a77be0485142291c9c1ec006aac3d018189240bc4f94c8f175b4cf291b8f" exitCode=2 Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.377579 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-t6d75" event={"ID":"42c712ed-b850-4947-b16e-5517c26612f8","Type":"ContainerDied","Data":"c4d2a77be0485142291c9c1ec006aac3d018189240bc4f94c8f175b4cf291b8f"} Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.377936 4696 scope.go:117] "RemoveContainer" containerID="c4d2a77be0485142291c9c1ec006aac3d018189240bc4f94c8f175b4cf291b8f" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.405427 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9tp76"] Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.408333 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9tp76"] Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.416137 4696 scope.go:117] "RemoveContainer" containerID="7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.447968 4696 scope.go:117] "RemoveContainer" containerID="4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.479082 4696 scope.go:117] "RemoveContainer" containerID="beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.508287 4696 scope.go:117] "RemoveContainer" containerID="2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.531215 4696 scope.go:117] "RemoveContainer" containerID="cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.545588 4696 scope.go:117] "RemoveContainer" containerID="92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.563433 4696 scope.go:117] "RemoveContainer" containerID="42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.628755 4696 scope.go:117] "RemoveContainer" containerID="af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.653037 4696 scope.go:117] "RemoveContainer" containerID="4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281" Mar 21 08:38:45 crc kubenswrapper[4696]: E0321 08:38:45.653687 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281\": container with ID starting with 4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281 not found: ID does not exist" containerID="4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.653716 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281"} err="failed to get container status \"4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281\": rpc error: code = NotFound desc = could not find container \"4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281\": container with ID starting with 4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.653738 4696 scope.go:117] "RemoveContainer" containerID="7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0" Mar 21 08:38:45 crc kubenswrapper[4696]: E0321 08:38:45.654051 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0\": container with ID starting with 7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0 not found: ID does not exist" containerID="7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.654091 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0"} err="failed to get container status \"7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0\": rpc error: code = NotFound desc = could not find container \"7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0\": container with ID starting with 7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.654116 4696 scope.go:117] "RemoveContainer" containerID="4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8" Mar 21 08:38:45 crc kubenswrapper[4696]: E0321 08:38:45.654434 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8\": container with ID starting with 4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8 not found: ID does not exist" containerID="4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.654471 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8"} err="failed to get container status \"4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8\": rpc error: code = NotFound desc = could not find container \"4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8\": container with ID starting with 4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.654495 4696 scope.go:117] "RemoveContainer" containerID="beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a" Mar 21 08:38:45 crc kubenswrapper[4696]: E0321 08:38:45.654799 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a\": container with ID starting with beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a not found: ID does not exist" containerID="beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.654839 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a"} err="failed to get container status \"beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a\": rpc error: code = NotFound desc = could not find container \"beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a\": container with ID starting with beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.654854 4696 scope.go:117] "RemoveContainer" containerID="2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a" Mar 21 08:38:45 crc kubenswrapper[4696]: E0321 08:38:45.655132 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a\": container with ID starting with 2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a not found: ID does not exist" containerID="2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.655156 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a"} err="failed to get container status \"2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a\": rpc error: code = NotFound desc = could not find container \"2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a\": container with ID starting with 2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.655195 4696 scope.go:117] "RemoveContainer" containerID="cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7" Mar 21 08:38:45 crc kubenswrapper[4696]: E0321 08:38:45.655450 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7\": container with ID starting with cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7 not found: ID does not exist" containerID="cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.655472 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7"} err="failed to get container status \"cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7\": rpc error: code = NotFound desc = could not find container \"cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7\": container with ID starting with cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.655489 4696 scope.go:117] "RemoveContainer" containerID="92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59" Mar 21 08:38:45 crc kubenswrapper[4696]: E0321 08:38:45.655700 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59\": container with ID starting with 92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59 not found: ID does not exist" containerID="92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.655719 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59"} err="failed to get container status \"92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59\": rpc error: code = NotFound desc = could not find container \"92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59\": container with ID starting with 92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.655731 4696 scope.go:117] "RemoveContainer" containerID="42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054" Mar 21 08:38:45 crc kubenswrapper[4696]: E0321 08:38:45.655985 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054\": container with ID starting with 42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054 not found: ID does not exist" containerID="42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.656003 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054"} err="failed to get container status \"42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054\": rpc error: code = NotFound desc = could not find container \"42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054\": container with ID starting with 42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.656015 4696 scope.go:117] "RemoveContainer" containerID="af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c" Mar 21 08:38:45 crc kubenswrapper[4696]: E0321 08:38:45.656281 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c\": container with ID starting with af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c not found: ID does not exist" containerID="af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.656303 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c"} err="failed to get container status \"af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c\": rpc error: code = NotFound desc = could not find container \"af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c\": container with ID starting with af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.656320 4696 scope.go:117] "RemoveContainer" containerID="4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.658228 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281"} err="failed to get container status \"4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281\": rpc error: code = NotFound desc = could not find container \"4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281\": container with ID starting with 4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.658247 4696 scope.go:117] "RemoveContainer" containerID="7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.658543 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0"} err="failed to get container status \"7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0\": rpc error: code = NotFound desc = could not find container \"7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0\": container with ID starting with 7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.658572 4696 scope.go:117] "RemoveContainer" containerID="4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.659048 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8"} err="failed to get container status \"4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8\": rpc error: code = NotFound desc = could not find container \"4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8\": container with ID starting with 4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.659068 4696 scope.go:117] "RemoveContainer" containerID="beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.659342 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a"} err="failed to get container status \"beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a\": rpc error: code = NotFound desc = could not find container \"beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a\": container with ID starting with beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.659357 4696 scope.go:117] "RemoveContainer" containerID="2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.659680 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a"} err="failed to get container status \"2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a\": rpc error: code = NotFound desc = could not find container \"2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a\": container with ID starting with 2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.659705 4696 scope.go:117] "RemoveContainer" containerID="cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.663956 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7"} err="failed to get container status \"cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7\": rpc error: code = NotFound desc = could not find container \"cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7\": container with ID starting with cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.663984 4696 scope.go:117] "RemoveContainer" containerID="92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.664375 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59"} err="failed to get container status \"92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59\": rpc error: code = NotFound desc = could not find container \"92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59\": container with ID starting with 92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.664403 4696 scope.go:117] "RemoveContainer" containerID="42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.664850 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054"} err="failed to get container status \"42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054\": rpc error: code = NotFound desc = could not find container \"42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054\": container with ID starting with 42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.664873 4696 scope.go:117] "RemoveContainer" containerID="af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.665126 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c"} err="failed to get container status \"af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c\": rpc error: code = NotFound desc = could not find container \"af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c\": container with ID starting with af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.665148 4696 scope.go:117] "RemoveContainer" containerID="4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.665377 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281"} err="failed to get container status \"4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281\": rpc error: code = NotFound desc = could not find container \"4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281\": container with ID starting with 4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.665393 4696 scope.go:117] "RemoveContainer" containerID="7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.665634 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0"} err="failed to get container status \"7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0\": rpc error: code = NotFound desc = could not find container \"7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0\": container with ID starting with 7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.665648 4696 scope.go:117] "RemoveContainer" containerID="4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.665901 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8"} err="failed to get container status \"4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8\": rpc error: code = NotFound desc = could not find container \"4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8\": container with ID starting with 4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.665922 4696 scope.go:117] "RemoveContainer" containerID="beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.666183 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a"} err="failed to get container status \"beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a\": rpc error: code = NotFound desc = could not find container \"beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a\": container with ID starting with beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.666205 4696 scope.go:117] "RemoveContainer" containerID="2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.666454 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a"} err="failed to get container status \"2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a\": rpc error: code = NotFound desc = could not find container \"2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a\": container with ID starting with 2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.666475 4696 scope.go:117] "RemoveContainer" containerID="cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.666720 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7"} err="failed to get container status \"cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7\": rpc error: code = NotFound desc = could not find container \"cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7\": container with ID starting with cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.666740 4696 scope.go:117] "RemoveContainer" containerID="92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.667019 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59"} err="failed to get container status \"92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59\": rpc error: code = NotFound desc = could not find container \"92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59\": container with ID starting with 92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.667040 4696 scope.go:117] "RemoveContainer" containerID="42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.667284 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054"} err="failed to get container status \"42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054\": rpc error: code = NotFound desc = could not find container \"42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054\": container with ID starting with 42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.667301 4696 scope.go:117] "RemoveContainer" containerID="af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.667542 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c"} err="failed to get container status \"af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c\": rpc error: code = NotFound desc = could not find container \"af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c\": container with ID starting with af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.667564 4696 scope.go:117] "RemoveContainer" containerID="4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.667807 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281"} err="failed to get container status \"4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281\": rpc error: code = NotFound desc = could not find container \"4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281\": container with ID starting with 4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.667887 4696 scope.go:117] "RemoveContainer" containerID="7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.668159 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0"} err="failed to get container status \"7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0\": rpc error: code = NotFound desc = could not find container \"7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0\": container with ID starting with 7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.668180 4696 scope.go:117] "RemoveContainer" containerID="4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.668435 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8"} err="failed to get container status \"4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8\": rpc error: code = NotFound desc = could not find container \"4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8\": container with ID starting with 4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.668459 4696 scope.go:117] "RemoveContainer" containerID="beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.668711 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a"} err="failed to get container status \"beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a\": rpc error: code = NotFound desc = could not find container \"beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a\": container with ID starting with beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.668737 4696 scope.go:117] "RemoveContainer" containerID="2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.668999 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a"} err="failed to get container status \"2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a\": rpc error: code = NotFound desc = could not find container \"2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a\": container with ID starting with 2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.669020 4696 scope.go:117] "RemoveContainer" containerID="cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.669274 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7"} err="failed to get container status \"cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7\": rpc error: code = NotFound desc = could not find container \"cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7\": container with ID starting with cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.669295 4696 scope.go:117] "RemoveContainer" containerID="92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.669538 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59"} err="failed to get container status \"92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59\": rpc error: code = NotFound desc = could not find container \"92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59\": container with ID starting with 92e0506eb9c96bf266e5b5763f9e582f2ff540fe1ce9bb54aa8643bb27695d59 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.669560 4696 scope.go:117] "RemoveContainer" containerID="42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.669808 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054"} err="failed to get container status \"42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054\": rpc error: code = NotFound desc = could not find container \"42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054\": container with ID starting with 42c7bc4b3178d3535f9985b95871261471f4fd4f4bd5fda891188e1cea48c054 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.669849 4696 scope.go:117] "RemoveContainer" containerID="af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.670123 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c"} err="failed to get container status \"af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c\": rpc error: code = NotFound desc = could not find container \"af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c\": container with ID starting with af297945784e50c051377b53f39106b23b877dc51319d9ead947d4c2f4ca480c not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.670144 4696 scope.go:117] "RemoveContainer" containerID="4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.670385 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281"} err="failed to get container status \"4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281\": rpc error: code = NotFound desc = could not find container \"4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281\": container with ID starting with 4efd7e49f0ae88592c5d10f0b25d6deeeb9eb0cbe892c2b53806707c8d6f8281 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.670406 4696 scope.go:117] "RemoveContainer" containerID="7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.670670 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0"} err="failed to get container status \"7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0\": rpc error: code = NotFound desc = could not find container \"7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0\": container with ID starting with 7ffc7705492da7e2c603aaa19fb733e73b42516675ccc1d388bace0cace014e0 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.670756 4696 scope.go:117] "RemoveContainer" containerID="4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.672492 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8"} err="failed to get container status \"4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8\": rpc error: code = NotFound desc = could not find container \"4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8\": container with ID starting with 4c04f106ebd4f6cf655315f852aa123600602415472821fc3b00a6c532c176a8 not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.672575 4696 scope.go:117] "RemoveContainer" containerID="beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.672892 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a"} err="failed to get container status \"beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a\": rpc error: code = NotFound desc = could not find container \"beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a\": container with ID starting with beecf016b9ae6c91a551199717fce1413e057f3680b5aa62032b66b81743932a not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.672960 4696 scope.go:117] "RemoveContainer" containerID="2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.673339 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a"} err="failed to get container status \"2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a\": rpc error: code = NotFound desc = could not find container \"2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a\": container with ID starting with 2ad450fceea48efdba22eaea7327a96456eacbc70c94cee36b8d3f49ccee051a not found: ID does not exist" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.673405 4696 scope.go:117] "RemoveContainer" containerID="cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7" Mar 21 08:38:45 crc kubenswrapper[4696]: I0321 08:38:45.673637 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7"} err="failed to get container status \"cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7\": rpc error: code = NotFound desc = could not find container \"cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7\": container with ID starting with cde70b7256283e60c2c4168fe971e340fce2ad1ce831ce83b7392b4420adb8b7 not found: ID does not exist" Mar 21 08:38:46 crc kubenswrapper[4696]: I0321 08:38:46.384004 4696 generic.go:334] "Generic (PLEG): container finished" podID="d53ef7be-e64a-4164-b951-03979405c91f" containerID="eefd70b22ef10d872c7d5c6f2959cd20b16eb1ea55bd78b6aae261977689e7b1" exitCode=0 Mar 21 08:38:46 crc kubenswrapper[4696]: I0321 08:38:46.384236 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" event={"ID":"d53ef7be-e64a-4164-b951-03979405c91f","Type":"ContainerDied","Data":"eefd70b22ef10d872c7d5c6f2959cd20b16eb1ea55bd78b6aae261977689e7b1"} Mar 21 08:38:46 crc kubenswrapper[4696]: I0321 08:38:46.385217 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" event={"ID":"d53ef7be-e64a-4164-b951-03979405c91f","Type":"ContainerStarted","Data":"9ee45c48c08498c51221780230e50a28570c2224c2b7cd420719747511e2af3e"} Mar 21 08:38:46 crc kubenswrapper[4696]: I0321 08:38:46.386740 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-t6d75_42c712ed-b850-4947-b16e-5517c26612f8/kube-multus/0.log" Mar 21 08:38:46 crc kubenswrapper[4696]: I0321 08:38:46.386786 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-t6d75" event={"ID":"42c712ed-b850-4947-b16e-5517c26612f8","Type":"ContainerStarted","Data":"0c07a2addfbdbbd37caeb3e2b85bc3627465a92b321df2436f0e6c010b2247cc"} Mar 21 08:38:46 crc kubenswrapper[4696]: I0321 08:38:46.548052 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f0e7707-cb43-4f15-b309-7f651da35ddf" path="/var/lib/kubelet/pods/9f0e7707-cb43-4f15-b309-7f651da35ddf/volumes" Mar 21 08:38:46 crc kubenswrapper[4696]: I0321 08:38:46.838689 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-8ff7d675-rs27t"] Mar 21 08:38:46 crc kubenswrapper[4696]: I0321 08:38:46.839615 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-8ff7d675-rs27t" Mar 21 08:38:46 crc kubenswrapper[4696]: I0321 08:38:46.841913 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-qqmft" Mar 21 08:38:46 crc kubenswrapper[4696]: I0321 08:38:46.841964 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Mar 21 08:38:46 crc kubenswrapper[4696]: I0321 08:38:46.843730 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Mar 21 08:38:46 crc kubenswrapper[4696]: I0321 08:38:46.867071 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqcks\" (UniqueName: \"kubernetes.io/projected/527ae575-3dfd-4f0a-9af9-63668af676d3-kube-api-access-hqcks\") pod \"obo-prometheus-operator-8ff7d675-rs27t\" (UID: \"527ae575-3dfd-4f0a-9af9-63668af676d3\") " pod="openshift-operators/obo-prometheus-operator-8ff7d675-rs27t" Mar 21 08:38:46 crc kubenswrapper[4696]: I0321 08:38:46.968524 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqcks\" (UniqueName: \"kubernetes.io/projected/527ae575-3dfd-4f0a-9af9-63668af676d3-kube-api-access-hqcks\") pod \"obo-prometheus-operator-8ff7d675-rs27t\" (UID: \"527ae575-3dfd-4f0a-9af9-63668af676d3\") " pod="openshift-operators/obo-prometheus-operator-8ff7d675-rs27t" Mar 21 08:38:46 crc kubenswrapper[4696]: I0321 08:38:46.975401 4696 scope.go:117] "RemoveContainer" containerID="4a956d1c59ecd49c327f9b814d7d2c78baf1cb3ff0693847794badb2a86f9dd7" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.022070 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqcks\" (UniqueName: \"kubernetes.io/projected/527ae575-3dfd-4f0a-9af9-63668af676d3-kube-api-access-hqcks\") pod \"obo-prometheus-operator-8ff7d675-rs27t\" (UID: \"527ae575-3dfd-4f0a-9af9-63668af676d3\") " pod="openshift-operators/obo-prometheus-operator-8ff7d675-rs27t" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.155234 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-8ff7d675-rs27t" Mar 21 08:38:47 crc kubenswrapper[4696]: E0321 08:38:47.175343 4696 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-8ff7d675-rs27t_openshift-operators_527ae575-3dfd-4f0a-9af9-63668af676d3_0(f7ef4613ff7925ad89bea3d95edab13b862bc84d0c69989782152990aa0fa2fe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 08:38:47 crc kubenswrapper[4696]: E0321 08:38:47.175424 4696 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-8ff7d675-rs27t_openshift-operators_527ae575-3dfd-4f0a-9af9-63668af676d3_0(f7ef4613ff7925ad89bea3d95edab13b862bc84d0c69989782152990aa0fa2fe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-8ff7d675-rs27t" Mar 21 08:38:47 crc kubenswrapper[4696]: E0321 08:38:47.175449 4696 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-8ff7d675-rs27t_openshift-operators_527ae575-3dfd-4f0a-9af9-63668af676d3_0(f7ef4613ff7925ad89bea3d95edab13b862bc84d0c69989782152990aa0fa2fe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-8ff7d675-rs27t" Mar 21 08:38:47 crc kubenswrapper[4696]: E0321 08:38:47.175500 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-8ff7d675-rs27t_openshift-operators(527ae575-3dfd-4f0a-9af9-63668af676d3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-8ff7d675-rs27t_openshift-operators(527ae575-3dfd-4f0a-9af9-63668af676d3)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-8ff7d675-rs27t_openshift-operators_527ae575-3dfd-4f0a-9af9-63668af676d3_0(f7ef4613ff7925ad89bea3d95edab13b862bc84d0c69989782152990aa0fa2fe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-8ff7d675-rs27t" podUID="527ae575-3dfd-4f0a-9af9-63668af676d3" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.280140 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89"] Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.280954 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.283315 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-n9c68" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.283453 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.288988 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn"] Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.289855 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.373026 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d99a53b9-d503-4410-9c7c-807fca5276bc-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-db48b98cb-l8q89\" (UID: \"d99a53b9-d503-4410-9c7c-807fca5276bc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.373421 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/474e7a66-fb1d-448c-8297-b511ec7b6e7c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-db48b98cb-78qsn\" (UID: \"474e7a66-fb1d-448c-8297-b511ec7b6e7c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.373491 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d99a53b9-d503-4410-9c7c-807fca5276bc-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-db48b98cb-l8q89\" (UID: \"d99a53b9-d503-4410-9c7c-807fca5276bc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.373596 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/474e7a66-fb1d-448c-8297-b511ec7b6e7c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-db48b98cb-78qsn\" (UID: \"474e7a66-fb1d-448c-8297-b511ec7b6e7c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.395834 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" event={"ID":"d53ef7be-e64a-4164-b951-03979405c91f","Type":"ContainerStarted","Data":"c5f0d58339998f27646322d02b5a531d5ff697fe12b161835acd893c24e201ed"} Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.395872 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" event={"ID":"d53ef7be-e64a-4164-b951-03979405c91f","Type":"ContainerStarted","Data":"1ab13997baee1b0ea58db4b139617f60fe8651f74d4d0cad4c37fbe194da2b7a"} Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.395883 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" event={"ID":"d53ef7be-e64a-4164-b951-03979405c91f","Type":"ContainerStarted","Data":"91719704e1fd568daaf618bc203613d7f9e8e9aa5a75025bc027a66cf3a0c87d"} Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.395894 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" event={"ID":"d53ef7be-e64a-4164-b951-03979405c91f","Type":"ContainerStarted","Data":"60fa10208aea5c6c67e45dde21610d58eb18cadaf602c31637946669925f205b"} Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.395902 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" event={"ID":"d53ef7be-e64a-4164-b951-03979405c91f","Type":"ContainerStarted","Data":"98e13121e3db93a20bb84f38d8d7df771bf1ad319d4f839f0b7c42d2d030b43e"} Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.395910 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" event={"ID":"d53ef7be-e64a-4164-b951-03979405c91f","Type":"ContainerStarted","Data":"975016481e5f4b4d4ff0ce783d0408775a641254c159a902ba1eccaf969601d9"} Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.474946 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/474e7a66-fb1d-448c-8297-b511ec7b6e7c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-db48b98cb-78qsn\" (UID: \"474e7a66-fb1d-448c-8297-b511ec7b6e7c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.474994 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d99a53b9-d503-4410-9c7c-807fca5276bc-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-db48b98cb-l8q89\" (UID: \"d99a53b9-d503-4410-9c7c-807fca5276bc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.475027 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/474e7a66-fb1d-448c-8297-b511ec7b6e7c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-db48b98cb-78qsn\" (UID: \"474e7a66-fb1d-448c-8297-b511ec7b6e7c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.475068 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d99a53b9-d503-4410-9c7c-807fca5276bc-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-db48b98cb-l8q89\" (UID: \"d99a53b9-d503-4410-9c7c-807fca5276bc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.481270 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/474e7a66-fb1d-448c-8297-b511ec7b6e7c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-db48b98cb-78qsn\" (UID: \"474e7a66-fb1d-448c-8297-b511ec7b6e7c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.481292 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d99a53b9-d503-4410-9c7c-807fca5276bc-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-db48b98cb-l8q89\" (UID: \"d99a53b9-d503-4410-9c7c-807fca5276bc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.481271 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d99a53b9-d503-4410-9c7c-807fca5276bc-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-db48b98cb-l8q89\" (UID: \"d99a53b9-d503-4410-9c7c-807fca5276bc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.481639 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/474e7a66-fb1d-448c-8297-b511ec7b6e7c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-db48b98cb-78qsn\" (UID: \"474e7a66-fb1d-448c-8297-b511ec7b6e7c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.598697 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-6dd7dd855f-sp4gg"] Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.599325 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.601081 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-zzpm5" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.601100 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.604220 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.610635 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn" Mar 21 08:38:47 crc kubenswrapper[4696]: E0321 08:38:47.639227 4696 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-db48b98cb-l8q89_openshift-operators_d99a53b9-d503-4410-9c7c-807fca5276bc_0(5d4de69bfaae3c1b86ba22ef1b8a85c0690da13b4868691ab7c7500bbe9d132c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 08:38:47 crc kubenswrapper[4696]: E0321 08:38:47.639291 4696 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-db48b98cb-l8q89_openshift-operators_d99a53b9-d503-4410-9c7c-807fca5276bc_0(5d4de69bfaae3c1b86ba22ef1b8a85c0690da13b4868691ab7c7500bbe9d132c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89" Mar 21 08:38:47 crc kubenswrapper[4696]: E0321 08:38:47.639311 4696 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-db48b98cb-l8q89_openshift-operators_d99a53b9-d503-4410-9c7c-807fca5276bc_0(5d4de69bfaae3c1b86ba22ef1b8a85c0690da13b4868691ab7c7500bbe9d132c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89" Mar 21 08:38:47 crc kubenswrapper[4696]: E0321 08:38:47.639360 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-db48b98cb-l8q89_openshift-operators(d99a53b9-d503-4410-9c7c-807fca5276bc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-db48b98cb-l8q89_openshift-operators(d99a53b9-d503-4410-9c7c-807fca5276bc)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-db48b98cb-l8q89_openshift-operators_d99a53b9-d503-4410-9c7c-807fca5276bc_0(5d4de69bfaae3c1b86ba22ef1b8a85c0690da13b4868691ab7c7500bbe9d132c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89" podUID="d99a53b9-d503-4410-9c7c-807fca5276bc" Mar 21 08:38:47 crc kubenswrapper[4696]: E0321 08:38:47.645910 4696 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-db48b98cb-78qsn_openshift-operators_474e7a66-fb1d-448c-8297-b511ec7b6e7c_0(5d944cecba6cad98e8501f4c8a86856f9b8e9044701377ed2ed2cfa5507c9e32): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 08:38:47 crc kubenswrapper[4696]: E0321 08:38:47.646012 4696 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-db48b98cb-78qsn_openshift-operators_474e7a66-fb1d-448c-8297-b511ec7b6e7c_0(5d944cecba6cad98e8501f4c8a86856f9b8e9044701377ed2ed2cfa5507c9e32): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn" Mar 21 08:38:47 crc kubenswrapper[4696]: E0321 08:38:47.646035 4696 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-db48b98cb-78qsn_openshift-operators_474e7a66-fb1d-448c-8297-b511ec7b6e7c_0(5d944cecba6cad98e8501f4c8a86856f9b8e9044701377ed2ed2cfa5507c9e32): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn" Mar 21 08:38:47 crc kubenswrapper[4696]: E0321 08:38:47.646085 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-db48b98cb-78qsn_openshift-operators(474e7a66-fb1d-448c-8297-b511ec7b6e7c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-db48b98cb-78qsn_openshift-operators(474e7a66-fb1d-448c-8297-b511ec7b6e7c)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-db48b98cb-78qsn_openshift-operators_474e7a66-fb1d-448c-8297-b511ec7b6e7c_0(5d944cecba6cad98e8501f4c8a86856f9b8e9044701377ed2ed2cfa5507c9e32): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn" podUID="474e7a66-fb1d-448c-8297-b511ec7b6e7c" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.677276 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/e864a9cc-e270-4238-841b-77d0abad085d-observability-operator-tls\") pod \"observability-operator-6dd7dd855f-sp4gg\" (UID: \"e864a9cc-e270-4238-841b-77d0abad085d\") " pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.677325 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dngct\" (UniqueName: \"kubernetes.io/projected/e864a9cc-e270-4238-841b-77d0abad085d-kube-api-access-dngct\") pod \"observability-operator-6dd7dd855f-sp4gg\" (UID: \"e864a9cc-e270-4238-841b-77d0abad085d\") " pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.778759 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/e864a9cc-e270-4238-841b-77d0abad085d-observability-operator-tls\") pod \"observability-operator-6dd7dd855f-sp4gg\" (UID: \"e864a9cc-e270-4238-841b-77d0abad085d\") " pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.778830 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dngct\" (UniqueName: \"kubernetes.io/projected/e864a9cc-e270-4238-841b-77d0abad085d-kube-api-access-dngct\") pod \"observability-operator-6dd7dd855f-sp4gg\" (UID: \"e864a9cc-e270-4238-841b-77d0abad085d\") " pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.784002 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/e864a9cc-e270-4238-841b-77d0abad085d-observability-operator-tls\") pod \"observability-operator-6dd7dd855f-sp4gg\" (UID: \"e864a9cc-e270-4238-841b-77d0abad085d\") " pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.798557 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dngct\" (UniqueName: \"kubernetes.io/projected/e864a9cc-e270-4238-841b-77d0abad085d-kube-api-access-dngct\") pod \"observability-operator-6dd7dd855f-sp4gg\" (UID: \"e864a9cc-e270-4238-841b-77d0abad085d\") " pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" Mar 21 08:38:47 crc kubenswrapper[4696]: I0321 08:38:47.917165 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" Mar 21 08:38:47 crc kubenswrapper[4696]: E0321 08:38:47.936028 4696 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-6dd7dd855f-sp4gg_openshift-operators_e864a9cc-e270-4238-841b-77d0abad085d_0(9510655f0d9ffed304b7e8833845db1dd71b333b52f14b514d9a9af25b377376): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 08:38:47 crc kubenswrapper[4696]: E0321 08:38:47.936099 4696 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-6dd7dd855f-sp4gg_openshift-operators_e864a9cc-e270-4238-841b-77d0abad085d_0(9510655f0d9ffed304b7e8833845db1dd71b333b52f14b514d9a9af25b377376): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" Mar 21 08:38:47 crc kubenswrapper[4696]: E0321 08:38:47.936123 4696 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-6dd7dd855f-sp4gg_openshift-operators_e864a9cc-e270-4238-841b-77d0abad085d_0(9510655f0d9ffed304b7e8833845db1dd71b333b52f14b514d9a9af25b377376): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" Mar 21 08:38:47 crc kubenswrapper[4696]: E0321 08:38:47.936188 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-6dd7dd855f-sp4gg_openshift-operators(e864a9cc-e270-4238-841b-77d0abad085d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-6dd7dd855f-sp4gg_openshift-operators(e864a9cc-e270-4238-841b-77d0abad085d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-6dd7dd855f-sp4gg_openshift-operators_e864a9cc-e270-4238-841b-77d0abad085d_0(9510655f0d9ffed304b7e8833845db1dd71b333b52f14b514d9a9af25b377376): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" podUID="e864a9cc-e270-4238-841b-77d0abad085d" Mar 21 08:38:48 crc kubenswrapper[4696]: I0321 08:38:48.001069 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-64bcb4cf6f-jjk5b"] Mar 21 08:38:48 crc kubenswrapper[4696]: I0321 08:38:48.001732 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:38:48 crc kubenswrapper[4696]: I0321 08:38:48.004501 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-service-cert" Mar 21 08:38:48 crc kubenswrapper[4696]: I0321 08:38:48.005008 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-bt968" Mar 21 08:38:48 crc kubenswrapper[4696]: I0321 08:38:48.081275 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/127ee197-aa20-41d3-ac26-e8aead566b64-apiservice-cert\") pod \"perses-operator-64bcb4cf6f-jjk5b\" (UID: \"127ee197-aa20-41d3-ac26-e8aead566b64\") " pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:38:48 crc kubenswrapper[4696]: I0321 08:38:48.081550 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sqx4\" (UniqueName: \"kubernetes.io/projected/127ee197-aa20-41d3-ac26-e8aead566b64-kube-api-access-7sqx4\") pod \"perses-operator-64bcb4cf6f-jjk5b\" (UID: \"127ee197-aa20-41d3-ac26-e8aead566b64\") " pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:38:48 crc kubenswrapper[4696]: I0321 08:38:48.081641 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/127ee197-aa20-41d3-ac26-e8aead566b64-openshift-service-ca\") pod \"perses-operator-64bcb4cf6f-jjk5b\" (UID: \"127ee197-aa20-41d3-ac26-e8aead566b64\") " pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:38:48 crc kubenswrapper[4696]: I0321 08:38:48.081719 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/127ee197-aa20-41d3-ac26-e8aead566b64-webhook-cert\") pod \"perses-operator-64bcb4cf6f-jjk5b\" (UID: \"127ee197-aa20-41d3-ac26-e8aead566b64\") " pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:38:48 crc kubenswrapper[4696]: I0321 08:38:48.183179 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/127ee197-aa20-41d3-ac26-e8aead566b64-openshift-service-ca\") pod \"perses-operator-64bcb4cf6f-jjk5b\" (UID: \"127ee197-aa20-41d3-ac26-e8aead566b64\") " pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:38:48 crc kubenswrapper[4696]: I0321 08:38:48.183503 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/127ee197-aa20-41d3-ac26-e8aead566b64-webhook-cert\") pod \"perses-operator-64bcb4cf6f-jjk5b\" (UID: \"127ee197-aa20-41d3-ac26-e8aead566b64\") " pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:38:48 crc kubenswrapper[4696]: I0321 08:38:48.184289 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/127ee197-aa20-41d3-ac26-e8aead566b64-apiservice-cert\") pod \"perses-operator-64bcb4cf6f-jjk5b\" (UID: \"127ee197-aa20-41d3-ac26-e8aead566b64\") " pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:38:48 crc kubenswrapper[4696]: I0321 08:38:48.184182 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/127ee197-aa20-41d3-ac26-e8aead566b64-openshift-service-ca\") pod \"perses-operator-64bcb4cf6f-jjk5b\" (UID: \"127ee197-aa20-41d3-ac26-e8aead566b64\") " pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:38:48 crc kubenswrapper[4696]: I0321 08:38:48.184478 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sqx4\" (UniqueName: \"kubernetes.io/projected/127ee197-aa20-41d3-ac26-e8aead566b64-kube-api-access-7sqx4\") pod \"perses-operator-64bcb4cf6f-jjk5b\" (UID: \"127ee197-aa20-41d3-ac26-e8aead566b64\") " pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:38:48 crc kubenswrapper[4696]: I0321 08:38:48.189412 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/127ee197-aa20-41d3-ac26-e8aead566b64-apiservice-cert\") pod \"perses-operator-64bcb4cf6f-jjk5b\" (UID: \"127ee197-aa20-41d3-ac26-e8aead566b64\") " pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:38:48 crc kubenswrapper[4696]: I0321 08:38:48.189546 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/127ee197-aa20-41d3-ac26-e8aead566b64-webhook-cert\") pod \"perses-operator-64bcb4cf6f-jjk5b\" (UID: \"127ee197-aa20-41d3-ac26-e8aead566b64\") " pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:38:48 crc kubenswrapper[4696]: I0321 08:38:48.202015 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sqx4\" (UniqueName: \"kubernetes.io/projected/127ee197-aa20-41d3-ac26-e8aead566b64-kube-api-access-7sqx4\") pod \"perses-operator-64bcb4cf6f-jjk5b\" (UID: \"127ee197-aa20-41d3-ac26-e8aead566b64\") " pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:38:48 crc kubenswrapper[4696]: I0321 08:38:48.315993 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:38:48 crc kubenswrapper[4696]: E0321 08:38:48.336606 4696 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-64bcb4cf6f-jjk5b_openshift-operators_127ee197-aa20-41d3-ac26-e8aead566b64_0(9cff564fbc09442109352c9efe05b469cde2248e667097623c3b96a067f65aa8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 08:38:48 crc kubenswrapper[4696]: E0321 08:38:48.336798 4696 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-64bcb4cf6f-jjk5b_openshift-operators_127ee197-aa20-41d3-ac26-e8aead566b64_0(9cff564fbc09442109352c9efe05b469cde2248e667097623c3b96a067f65aa8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:38:48 crc kubenswrapper[4696]: E0321 08:38:48.336948 4696 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-64bcb4cf6f-jjk5b_openshift-operators_127ee197-aa20-41d3-ac26-e8aead566b64_0(9cff564fbc09442109352c9efe05b469cde2248e667097623c3b96a067f65aa8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:38:48 crc kubenswrapper[4696]: E0321 08:38:48.337084 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-64bcb4cf6f-jjk5b_openshift-operators(127ee197-aa20-41d3-ac26-e8aead566b64)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-64bcb4cf6f-jjk5b_openshift-operators(127ee197-aa20-41d3-ac26-e8aead566b64)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-64bcb4cf6f-jjk5b_openshift-operators_127ee197-aa20-41d3-ac26-e8aead566b64_0(9cff564fbc09442109352c9efe05b469cde2248e667097623c3b96a067f65aa8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" podUID="127ee197-aa20-41d3-ac26-e8aead566b64" Mar 21 08:38:49 crc kubenswrapper[4696]: I0321 08:38:49.427011 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" event={"ID":"d53ef7be-e64a-4164-b951-03979405c91f","Type":"ContainerStarted","Data":"bd9dfc1c7810b9e504e9b988235b0bd1fc6bde13523fdd4974061a7f4dbcd639"} Mar 21 08:38:52 crc kubenswrapper[4696]: I0321 08:38:52.450793 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" event={"ID":"d53ef7be-e64a-4164-b951-03979405c91f","Type":"ContainerStarted","Data":"5b0ea9ccba2de9be685376c1f33886e21a85c2ccd2f21730a4937df21dcc7212"} Mar 21 08:38:52 crc kubenswrapper[4696]: I0321 08:38:52.451326 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:52 crc kubenswrapper[4696]: I0321 08:38:52.451401 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:52 crc kubenswrapper[4696]: I0321 08:38:52.451469 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:52 crc kubenswrapper[4696]: I0321 08:38:52.479154 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:52 crc kubenswrapper[4696]: I0321 08:38:52.483473 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:38:52 crc kubenswrapper[4696]: I0321 08:38:52.484455 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" podStartSLOduration=7.484438886 podStartE2EDuration="7.484438886s" podCreationTimestamp="2026-03-21 08:38:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:38:52.48307736 +0000 UTC m=+666.603958073" watchObservedRunningTime="2026-03-21 08:38:52.484438886 +0000 UTC m=+666.605319589" Mar 21 08:38:53 crc kubenswrapper[4696]: I0321 08:38:53.170635 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn"] Mar 21 08:38:53 crc kubenswrapper[4696]: I0321 08:38:53.170776 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn" Mar 21 08:38:53 crc kubenswrapper[4696]: I0321 08:38:53.171287 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn" Mar 21 08:38:53 crc kubenswrapper[4696]: I0321 08:38:53.174686 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-6dd7dd855f-sp4gg"] Mar 21 08:38:53 crc kubenswrapper[4696]: I0321 08:38:53.174850 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" Mar 21 08:38:53 crc kubenswrapper[4696]: I0321 08:38:53.175338 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" Mar 21 08:38:53 crc kubenswrapper[4696]: I0321 08:38:53.182474 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89"] Mar 21 08:38:53 crc kubenswrapper[4696]: I0321 08:38:53.182599 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89" Mar 21 08:38:53 crc kubenswrapper[4696]: I0321 08:38:53.183039 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89" Mar 21 08:38:53 crc kubenswrapper[4696]: I0321 08:38:53.186114 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-64bcb4cf6f-jjk5b"] Mar 21 08:38:53 crc kubenswrapper[4696]: I0321 08:38:53.186220 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:38:53 crc kubenswrapper[4696]: I0321 08:38:53.186702 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:38:53 crc kubenswrapper[4696]: I0321 08:38:53.218862 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-8ff7d675-rs27t"] Mar 21 08:38:53 crc kubenswrapper[4696]: I0321 08:38:53.219272 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-8ff7d675-rs27t" Mar 21 08:38:53 crc kubenswrapper[4696]: I0321 08:38:53.219690 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-8ff7d675-rs27t" Mar 21 08:38:53 crc kubenswrapper[4696]: E0321 08:38:53.235385 4696 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-db48b98cb-78qsn_openshift-operators_474e7a66-fb1d-448c-8297-b511ec7b6e7c_0(46d204816a233a928c832765adddf8b1be790890038a2c98b7cec9805bff8ebf): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 08:38:53 crc kubenswrapper[4696]: E0321 08:38:53.235466 4696 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-db48b98cb-78qsn_openshift-operators_474e7a66-fb1d-448c-8297-b511ec7b6e7c_0(46d204816a233a928c832765adddf8b1be790890038a2c98b7cec9805bff8ebf): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn" Mar 21 08:38:53 crc kubenswrapper[4696]: E0321 08:38:53.235494 4696 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-db48b98cb-78qsn_openshift-operators_474e7a66-fb1d-448c-8297-b511ec7b6e7c_0(46d204816a233a928c832765adddf8b1be790890038a2c98b7cec9805bff8ebf): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn" Mar 21 08:38:53 crc kubenswrapper[4696]: E0321 08:38:53.235551 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-db48b98cb-78qsn_openshift-operators(474e7a66-fb1d-448c-8297-b511ec7b6e7c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-db48b98cb-78qsn_openshift-operators(474e7a66-fb1d-448c-8297-b511ec7b6e7c)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-db48b98cb-78qsn_openshift-operators_474e7a66-fb1d-448c-8297-b511ec7b6e7c_0(46d204816a233a928c832765adddf8b1be790890038a2c98b7cec9805bff8ebf): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn" podUID="474e7a66-fb1d-448c-8297-b511ec7b6e7c" Mar 21 08:38:53 crc kubenswrapper[4696]: E0321 08:38:53.250321 4696 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-6dd7dd855f-sp4gg_openshift-operators_e864a9cc-e270-4238-841b-77d0abad085d_0(2803b9eab0a27295bea1f48e3d7220eab27d3f1ef6dd5b4f8b73ca78b21bdcda): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 08:38:53 crc kubenswrapper[4696]: E0321 08:38:53.250391 4696 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-6dd7dd855f-sp4gg_openshift-operators_e864a9cc-e270-4238-841b-77d0abad085d_0(2803b9eab0a27295bea1f48e3d7220eab27d3f1ef6dd5b4f8b73ca78b21bdcda): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" Mar 21 08:38:53 crc kubenswrapper[4696]: E0321 08:38:53.250423 4696 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-6dd7dd855f-sp4gg_openshift-operators_e864a9cc-e270-4238-841b-77d0abad085d_0(2803b9eab0a27295bea1f48e3d7220eab27d3f1ef6dd5b4f8b73ca78b21bdcda): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" Mar 21 08:38:53 crc kubenswrapper[4696]: E0321 08:38:53.250475 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-6dd7dd855f-sp4gg_openshift-operators(e864a9cc-e270-4238-841b-77d0abad085d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-6dd7dd855f-sp4gg_openshift-operators(e864a9cc-e270-4238-841b-77d0abad085d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-6dd7dd855f-sp4gg_openshift-operators_e864a9cc-e270-4238-841b-77d0abad085d_0(2803b9eab0a27295bea1f48e3d7220eab27d3f1ef6dd5b4f8b73ca78b21bdcda): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" podUID="e864a9cc-e270-4238-841b-77d0abad085d" Mar 21 08:38:53 crc kubenswrapper[4696]: E0321 08:38:53.268057 4696 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-db48b98cb-l8q89_openshift-operators_d99a53b9-d503-4410-9c7c-807fca5276bc_0(fe76ead422796403cd88e1e25cf277f2c8e6cd510620b372b6743b0a56ec81ee): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 08:38:53 crc kubenswrapper[4696]: E0321 08:38:53.268126 4696 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-db48b98cb-l8q89_openshift-operators_d99a53b9-d503-4410-9c7c-807fca5276bc_0(fe76ead422796403cd88e1e25cf277f2c8e6cd510620b372b6743b0a56ec81ee): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89" Mar 21 08:38:53 crc kubenswrapper[4696]: E0321 08:38:53.268170 4696 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-db48b98cb-l8q89_openshift-operators_d99a53b9-d503-4410-9c7c-807fca5276bc_0(fe76ead422796403cd88e1e25cf277f2c8e6cd510620b372b6743b0a56ec81ee): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89" Mar 21 08:38:53 crc kubenswrapper[4696]: E0321 08:38:53.268226 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-db48b98cb-l8q89_openshift-operators(d99a53b9-d503-4410-9c7c-807fca5276bc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-db48b98cb-l8q89_openshift-operators(d99a53b9-d503-4410-9c7c-807fca5276bc)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-db48b98cb-l8q89_openshift-operators_d99a53b9-d503-4410-9c7c-807fca5276bc_0(fe76ead422796403cd88e1e25cf277f2c8e6cd510620b372b6743b0a56ec81ee): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89" podUID="d99a53b9-d503-4410-9c7c-807fca5276bc" Mar 21 08:38:53 crc kubenswrapper[4696]: E0321 08:38:53.277082 4696 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-64bcb4cf6f-jjk5b_openshift-operators_127ee197-aa20-41d3-ac26-e8aead566b64_0(c1d1cfb7a5a5384917c126c996096c36781caf7f2194b61751d0a7cce27c6f8b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 08:38:53 crc kubenswrapper[4696]: E0321 08:38:53.277140 4696 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-64bcb4cf6f-jjk5b_openshift-operators_127ee197-aa20-41d3-ac26-e8aead566b64_0(c1d1cfb7a5a5384917c126c996096c36781caf7f2194b61751d0a7cce27c6f8b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:38:53 crc kubenswrapper[4696]: E0321 08:38:53.277163 4696 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-64bcb4cf6f-jjk5b_openshift-operators_127ee197-aa20-41d3-ac26-e8aead566b64_0(c1d1cfb7a5a5384917c126c996096c36781caf7f2194b61751d0a7cce27c6f8b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:38:53 crc kubenswrapper[4696]: E0321 08:38:53.277206 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-64bcb4cf6f-jjk5b_openshift-operators(127ee197-aa20-41d3-ac26-e8aead566b64)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-64bcb4cf6f-jjk5b_openshift-operators(127ee197-aa20-41d3-ac26-e8aead566b64)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-64bcb4cf6f-jjk5b_openshift-operators_127ee197-aa20-41d3-ac26-e8aead566b64_0(c1d1cfb7a5a5384917c126c996096c36781caf7f2194b61751d0a7cce27c6f8b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" podUID="127ee197-aa20-41d3-ac26-e8aead566b64" Mar 21 08:38:53 crc kubenswrapper[4696]: E0321 08:38:53.288291 4696 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-8ff7d675-rs27t_openshift-operators_527ae575-3dfd-4f0a-9af9-63668af676d3_0(1eac7bd0a7729e9e7df395e771eb806fdfd4a6e797d3ceae99c13b81e72ee782): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 21 08:38:53 crc kubenswrapper[4696]: E0321 08:38:53.288365 4696 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-8ff7d675-rs27t_openshift-operators_527ae575-3dfd-4f0a-9af9-63668af676d3_0(1eac7bd0a7729e9e7df395e771eb806fdfd4a6e797d3ceae99c13b81e72ee782): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-8ff7d675-rs27t" Mar 21 08:38:53 crc kubenswrapper[4696]: E0321 08:38:53.288388 4696 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-8ff7d675-rs27t_openshift-operators_527ae575-3dfd-4f0a-9af9-63668af676d3_0(1eac7bd0a7729e9e7df395e771eb806fdfd4a6e797d3ceae99c13b81e72ee782): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-8ff7d675-rs27t" Mar 21 08:38:53 crc kubenswrapper[4696]: E0321 08:38:53.288431 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-8ff7d675-rs27t_openshift-operators(527ae575-3dfd-4f0a-9af9-63668af676d3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-8ff7d675-rs27t_openshift-operators(527ae575-3dfd-4f0a-9af9-63668af676d3)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-8ff7d675-rs27t_openshift-operators_527ae575-3dfd-4f0a-9af9-63668af676d3_0(1eac7bd0a7729e9e7df395e771eb806fdfd4a6e797d3ceae99c13b81e72ee782): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-8ff7d675-rs27t" podUID="527ae575-3dfd-4f0a-9af9-63668af676d3" Mar 21 08:39:03 crc kubenswrapper[4696]: I0321 08:39:03.534614 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" Mar 21 08:39:03 crc kubenswrapper[4696]: I0321 08:39:03.535308 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" Mar 21 08:39:03 crc kubenswrapper[4696]: I0321 08:39:03.875231 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-6dd7dd855f-sp4gg"] Mar 21 08:39:04 crc kubenswrapper[4696]: I0321 08:39:04.506490 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" event={"ID":"e864a9cc-e270-4238-841b-77d0abad085d","Type":"ContainerStarted","Data":"51d6f1cee42335752517d063e97b18b1acf271d561dd44a79719a1639145bc97"} Mar 21 08:39:05 crc kubenswrapper[4696]: I0321 08:39:05.534536 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-8ff7d675-rs27t" Mar 21 08:39:05 crc kubenswrapper[4696]: I0321 08:39:05.534962 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-8ff7d675-rs27t" Mar 21 08:39:05 crc kubenswrapper[4696]: I0321 08:39:05.775167 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-8ff7d675-rs27t"] Mar 21 08:39:06 crc kubenswrapper[4696]: I0321 08:39:06.519989 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-8ff7d675-rs27t" event={"ID":"527ae575-3dfd-4f0a-9af9-63668af676d3","Type":"ContainerStarted","Data":"71c30e0ad374a734dad3515d7156e20cd37c1ad9c9b25e85817eafc184a26eec"} Mar 21 08:39:06 crc kubenswrapper[4696]: I0321 08:39:06.539962 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89" Mar 21 08:39:06 crc kubenswrapper[4696]: I0321 08:39:06.540047 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn" Mar 21 08:39:06 crc kubenswrapper[4696]: I0321 08:39:06.540542 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89" Mar 21 08:39:06 crc kubenswrapper[4696]: I0321 08:39:06.540685 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn" Mar 21 08:39:08 crc kubenswrapper[4696]: I0321 08:39:08.536793 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:39:08 crc kubenswrapper[4696]: I0321 08:39:08.540031 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:39:10 crc kubenswrapper[4696]: I0321 08:39:10.586416 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89"] Mar 21 08:39:10 crc kubenswrapper[4696]: I0321 08:39:10.702160 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn"] Mar 21 08:39:11 crc kubenswrapper[4696]: I0321 08:39:11.947590 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-64bcb4cf6f-jjk5b"] Mar 21 08:39:12 crc kubenswrapper[4696]: I0321 08:39:12.551640 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" event={"ID":"e864a9cc-e270-4238-841b-77d0abad085d","Type":"ContainerStarted","Data":"33a77e0fcecc314a90d778fa0eb19a79d7e5ccf63a116efc01a71a642990b08f"} Mar 21 08:39:12 crc kubenswrapper[4696]: I0321 08:39:12.551874 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" Mar 21 08:39:12 crc kubenswrapper[4696]: I0321 08:39:12.552838 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" event={"ID":"127ee197-aa20-41d3-ac26-e8aead566b64","Type":"ContainerStarted","Data":"099e7d9156d08b647667025a8c22f9f8046915ee4dcfde390c2ea1c99e3f8e04"} Mar 21 08:39:12 crc kubenswrapper[4696]: I0321 08:39:12.554757 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-8ff7d675-rs27t" event={"ID":"527ae575-3dfd-4f0a-9af9-63668af676d3","Type":"ContainerStarted","Data":"00a8fb364b235fcf1b1f12795fbb94ef104191874d5abaa6d5a336501227bac1"} Mar 21 08:39:12 crc kubenswrapper[4696]: I0321 08:39:12.555667 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89" event={"ID":"d99a53b9-d503-4410-9c7c-807fca5276bc","Type":"ContainerStarted","Data":"97099c97608a7c4153b87fed0d67d26f712283c31218ccf65f5b5cc412d2b296"} Mar 21 08:39:12 crc kubenswrapper[4696]: I0321 08:39:12.557090 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn" event={"ID":"474e7a66-fb1d-448c-8297-b511ec7b6e7c","Type":"ContainerStarted","Data":"707ad3a8b0ce89aef930057ae92170e4316e9c1ad8d4d71adf975ac8002853dd"} Mar 21 08:39:12 crc kubenswrapper[4696]: I0321 08:39:12.567532 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" Mar 21 08:39:12 crc kubenswrapper[4696]: I0321 08:39:12.567553 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-6dd7dd855f-sp4gg" podStartSLOduration=17.903356385 podStartE2EDuration="25.567537557s" podCreationTimestamp="2026-03-21 08:38:47 +0000 UTC" firstStartedPulling="2026-03-21 08:39:03.88239371 +0000 UTC m=+678.003274423" lastFinishedPulling="2026-03-21 08:39:11.546574882 +0000 UTC m=+685.667455595" observedRunningTime="2026-03-21 08:39:12.565113585 +0000 UTC m=+686.685994298" watchObservedRunningTime="2026-03-21 08:39:12.567537557 +0000 UTC m=+686.688418260" Mar 21 08:39:12 crc kubenswrapper[4696]: I0321 08:39:12.591762 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-8ff7d675-rs27t" podStartSLOduration=20.769874263 podStartE2EDuration="26.591738586s" podCreationTimestamp="2026-03-21 08:38:46 +0000 UTC" firstStartedPulling="2026-03-21 08:39:05.786304788 +0000 UTC m=+679.907185491" lastFinishedPulling="2026-03-21 08:39:11.608169101 +0000 UTC m=+685.729049814" observedRunningTime="2026-03-21 08:39:12.584511578 +0000 UTC m=+686.705392291" watchObservedRunningTime="2026-03-21 08:39:12.591738586 +0000 UTC m=+686.712619299" Mar 21 08:39:15 crc kubenswrapper[4696]: I0321 08:39:15.390462 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-klgqj" Mar 21 08:39:16 crc kubenswrapper[4696]: I0321 08:39:16.587244 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89" event={"ID":"d99a53b9-d503-4410-9c7c-807fca5276bc","Type":"ContainerStarted","Data":"9713b74f7ee2645648c7c4ad58dbf8a472fd3a366aa2d1cc965d5f9400c3d13c"} Mar 21 08:39:16 crc kubenswrapper[4696]: I0321 08:39:16.591138 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn" event={"ID":"474e7a66-fb1d-448c-8297-b511ec7b6e7c","Type":"ContainerStarted","Data":"78295ee636003c66c9eb2d9260acd0d592ada9dd70b717c47e44c041f4dd534e"} Mar 21 08:39:16 crc kubenswrapper[4696]: I0321 08:39:16.593058 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" event={"ID":"127ee197-aa20-41d3-ac26-e8aead566b64","Type":"ContainerStarted","Data":"695d82254593335192d0cae8388dec27c350f57706625845dc4f02dcdbe6e386"} Mar 21 08:39:16 crc kubenswrapper[4696]: I0321 08:39:16.593213 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:39:16 crc kubenswrapper[4696]: I0321 08:39:16.619851 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-78qsn" podStartSLOduration=25.0040917 podStartE2EDuration="29.61983609s" podCreationTimestamp="2026-03-21 08:38:47 +0000 UTC" firstStartedPulling="2026-03-21 08:39:11.594296881 +0000 UTC m=+685.715177594" lastFinishedPulling="2026-03-21 08:39:16.210041261 +0000 UTC m=+690.330921984" observedRunningTime="2026-03-21 08:39:16.616719549 +0000 UTC m=+690.737600262" watchObservedRunningTime="2026-03-21 08:39:16.61983609 +0000 UTC m=+690.740716803" Mar 21 08:39:16 crc kubenswrapper[4696]: I0321 08:39:16.646697 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" podStartSLOduration=25.394939618 podStartE2EDuration="29.646676457s" podCreationTimestamp="2026-03-21 08:38:47 +0000 UTC" firstStartedPulling="2026-03-21 08:39:11.963289551 +0000 UTC m=+686.084170264" lastFinishedPulling="2026-03-21 08:39:16.21502638 +0000 UTC m=+690.335907103" observedRunningTime="2026-03-21 08:39:16.644853479 +0000 UTC m=+690.765734212" watchObservedRunningTime="2026-03-21 08:39:16.646676457 +0000 UTC m=+690.767557170" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.282568 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-db48b98cb-l8q89" podStartSLOduration=29.682988699 podStartE2EDuration="34.282498878s" podCreationTimestamp="2026-03-21 08:38:47 +0000 UTC" firstStartedPulling="2026-03-21 08:39:11.544933809 +0000 UTC m=+685.665814522" lastFinishedPulling="2026-03-21 08:39:16.144443988 +0000 UTC m=+690.265324701" observedRunningTime="2026-03-21 08:39:16.667083126 +0000 UTC m=+690.787963829" watchObservedRunningTime="2026-03-21 08:39:21.282498878 +0000 UTC m=+695.403379591" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.283264 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-f85xz"] Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.283911 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-f85xz" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.290731 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w27t\" (UniqueName: \"kubernetes.io/projected/5a3acbc7-10ae-4287-9d73-15b97d7f0d71-kube-api-access-4w27t\") pod \"cert-manager-cainjector-cf98fcc89-f85xz\" (UID: \"5a3acbc7-10ae-4287-9d73-15b97d7f0d71\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-f85xz" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.292525 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.293195 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.305279 4696 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-jlwmr" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.306513 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-f85xz"] Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.312972 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-xbjnb"] Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.313952 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-xbjnb" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.324208 4696 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-7pt2j" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.330664 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-rp8tc"] Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.331567 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-rp8tc" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.333999 4696 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-czxh5" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.336315 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-xbjnb"] Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.349819 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-rp8tc"] Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.391509 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w27t\" (UniqueName: \"kubernetes.io/projected/5a3acbc7-10ae-4287-9d73-15b97d7f0d71-kube-api-access-4w27t\") pod \"cert-manager-cainjector-cf98fcc89-f85xz\" (UID: \"5a3acbc7-10ae-4287-9d73-15b97d7f0d71\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-f85xz" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.410475 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w27t\" (UniqueName: \"kubernetes.io/projected/5a3acbc7-10ae-4287-9d73-15b97d7f0d71-kube-api-access-4w27t\") pod \"cert-manager-cainjector-cf98fcc89-f85xz\" (UID: \"5a3acbc7-10ae-4287-9d73-15b97d7f0d71\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-f85xz" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.492992 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlm4x\" (UniqueName: \"kubernetes.io/projected/f6174e0b-3b88-4452-ab2f-89905f5d09a8-kube-api-access-zlm4x\") pod \"cert-manager-858654f9db-xbjnb\" (UID: \"f6174e0b-3b88-4452-ab2f-89905f5d09a8\") " pod="cert-manager/cert-manager-858654f9db-xbjnb" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.493124 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qqlr\" (UniqueName: \"kubernetes.io/projected/e0d6313d-e133-464b-b5ea-33699c8280c0-kube-api-access-8qqlr\") pod \"cert-manager-webhook-687f57d79b-rp8tc\" (UID: \"e0d6313d-e133-464b-b5ea-33699c8280c0\") " pod="cert-manager/cert-manager-webhook-687f57d79b-rp8tc" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.593991 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qqlr\" (UniqueName: \"kubernetes.io/projected/e0d6313d-e133-464b-b5ea-33699c8280c0-kube-api-access-8qqlr\") pod \"cert-manager-webhook-687f57d79b-rp8tc\" (UID: \"e0d6313d-e133-464b-b5ea-33699c8280c0\") " pod="cert-manager/cert-manager-webhook-687f57d79b-rp8tc" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.594072 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlm4x\" (UniqueName: \"kubernetes.io/projected/f6174e0b-3b88-4452-ab2f-89905f5d09a8-kube-api-access-zlm4x\") pod \"cert-manager-858654f9db-xbjnb\" (UID: \"f6174e0b-3b88-4452-ab2f-89905f5d09a8\") " pod="cert-manager/cert-manager-858654f9db-xbjnb" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.601811 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-f85xz" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.613961 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlm4x\" (UniqueName: \"kubernetes.io/projected/f6174e0b-3b88-4452-ab2f-89905f5d09a8-kube-api-access-zlm4x\") pod \"cert-manager-858654f9db-xbjnb\" (UID: \"f6174e0b-3b88-4452-ab2f-89905f5d09a8\") " pod="cert-manager/cert-manager-858654f9db-xbjnb" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.614037 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qqlr\" (UniqueName: \"kubernetes.io/projected/e0d6313d-e133-464b-b5ea-33699c8280c0-kube-api-access-8qqlr\") pod \"cert-manager-webhook-687f57d79b-rp8tc\" (UID: \"e0d6313d-e133-464b-b5ea-33699c8280c0\") " pod="cert-manager/cert-manager-webhook-687f57d79b-rp8tc" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.635047 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-xbjnb" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.657089 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-rp8tc" Mar 21 08:39:21 crc kubenswrapper[4696]: I0321 08:39:21.835938 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-f85xz"] Mar 21 08:39:21 crc kubenswrapper[4696]: W0321 08:39:21.858600 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a3acbc7_10ae_4287_9d73_15b97d7f0d71.slice/crio-e215f9e9df082383cac4ad5ce96d4f6a67304bb927f955a3355d951f1d6124fe WatchSource:0}: Error finding container e215f9e9df082383cac4ad5ce96d4f6a67304bb927f955a3355d951f1d6124fe: Status 404 returned error can't find the container with id e215f9e9df082383cac4ad5ce96d4f6a67304bb927f955a3355d951f1d6124fe Mar 21 08:39:22 crc kubenswrapper[4696]: W0321 08:39:22.099578 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6174e0b_3b88_4452_ab2f_89905f5d09a8.slice/crio-3574b30c3802bedb9146bf9803d8f6fc6e6aceb3964366af98cc6987d744613d WatchSource:0}: Error finding container 3574b30c3802bedb9146bf9803d8f6fc6e6aceb3964366af98cc6987d744613d: Status 404 returned error can't find the container with id 3574b30c3802bedb9146bf9803d8f6fc6e6aceb3964366af98cc6987d744613d Mar 21 08:39:22 crc kubenswrapper[4696]: I0321 08:39:22.101048 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-xbjnb"] Mar 21 08:39:22 crc kubenswrapper[4696]: W0321 08:39:22.102307 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0d6313d_e133_464b_b5ea_33699c8280c0.slice/crio-71f14e08d414c598589b17a0fb8b1d9bc1ec2ffee93932e749d7d4a0b2b3a28d WatchSource:0}: Error finding container 71f14e08d414c598589b17a0fb8b1d9bc1ec2ffee93932e749d7d4a0b2b3a28d: Status 404 returned error can't find the container with id 71f14e08d414c598589b17a0fb8b1d9bc1ec2ffee93932e749d7d4a0b2b3a28d Mar 21 08:39:22 crc kubenswrapper[4696]: I0321 08:39:22.107690 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-rp8tc"] Mar 21 08:39:22 crc kubenswrapper[4696]: I0321 08:39:22.630030 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-rp8tc" event={"ID":"e0d6313d-e133-464b-b5ea-33699c8280c0","Type":"ContainerStarted","Data":"71f14e08d414c598589b17a0fb8b1d9bc1ec2ffee93932e749d7d4a0b2b3a28d"} Mar 21 08:39:22 crc kubenswrapper[4696]: I0321 08:39:22.630994 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-xbjnb" event={"ID":"f6174e0b-3b88-4452-ab2f-89905f5d09a8","Type":"ContainerStarted","Data":"3574b30c3802bedb9146bf9803d8f6fc6e6aceb3964366af98cc6987d744613d"} Mar 21 08:39:22 crc kubenswrapper[4696]: I0321 08:39:22.632504 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-f85xz" event={"ID":"5a3acbc7-10ae-4287-9d73-15b97d7f0d71","Type":"ContainerStarted","Data":"e215f9e9df082383cac4ad5ce96d4f6a67304bb927f955a3355d951f1d6124fe"} Mar 21 08:39:24 crc kubenswrapper[4696]: I0321 08:39:24.654437 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-f85xz" event={"ID":"5a3acbc7-10ae-4287-9d73-15b97d7f0d71","Type":"ContainerStarted","Data":"498151d4d58d8f7565e7d8ee250743999d13ca640fb73e805ef16b9cbb969821"} Mar 21 08:39:24 crc kubenswrapper[4696]: I0321 08:39:24.677884 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-f85xz" podStartSLOduration=1.2261286359999999 podStartE2EDuration="3.677859786s" podCreationTimestamp="2026-03-21 08:39:21 +0000 UTC" firstStartedPulling="2026-03-21 08:39:21.861056528 +0000 UTC m=+695.981937231" lastFinishedPulling="2026-03-21 08:39:24.312787668 +0000 UTC m=+698.433668381" observedRunningTime="2026-03-21 08:39:24.668635036 +0000 UTC m=+698.789515749" watchObservedRunningTime="2026-03-21 08:39:24.677859786 +0000 UTC m=+698.798740499" Mar 21 08:39:26 crc kubenswrapper[4696]: I0321 08:39:26.664853 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-rp8tc" event={"ID":"e0d6313d-e133-464b-b5ea-33699c8280c0","Type":"ContainerStarted","Data":"6c9e152167f8c7ef2472855451f5911641bbf6032174ae41fefac46624445fe8"} Mar 21 08:39:26 crc kubenswrapper[4696]: I0321 08:39:26.665133 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-rp8tc" Mar 21 08:39:26 crc kubenswrapper[4696]: I0321 08:39:26.666284 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-xbjnb" event={"ID":"f6174e0b-3b88-4452-ab2f-89905f5d09a8","Type":"ContainerStarted","Data":"559aa03ff2bd67aad100e576b78bcbeac977aaae6fc4878e9f07cfd89a988bc5"} Mar 21 08:39:26 crc kubenswrapper[4696]: I0321 08:39:26.686948 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-rp8tc" podStartSLOduration=1.4853488559999999 podStartE2EDuration="5.686927894s" podCreationTimestamp="2026-03-21 08:39:21 +0000 UTC" firstStartedPulling="2026-03-21 08:39:22.104229541 +0000 UTC m=+696.225110254" lastFinishedPulling="2026-03-21 08:39:26.305808579 +0000 UTC m=+700.426689292" observedRunningTime="2026-03-21 08:39:26.685792144 +0000 UTC m=+700.806672867" watchObservedRunningTime="2026-03-21 08:39:26.686927894 +0000 UTC m=+700.807808607" Mar 21 08:39:26 crc kubenswrapper[4696]: I0321 08:39:26.704694 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-xbjnb" podStartSLOduration=1.5155830209999999 podStartE2EDuration="5.704671954s" podCreationTimestamp="2026-03-21 08:39:21 +0000 UTC" firstStartedPulling="2026-03-21 08:39:22.101313636 +0000 UTC m=+696.222194349" lastFinishedPulling="2026-03-21 08:39:26.290402559 +0000 UTC m=+700.411283282" observedRunningTime="2026-03-21 08:39:26.700837675 +0000 UTC m=+700.821718408" watchObservedRunningTime="2026-03-21 08:39:26.704671954 +0000 UTC m=+700.825552667" Mar 21 08:39:28 crc kubenswrapper[4696]: I0321 08:39:28.319388 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-64bcb4cf6f-jjk5b" Mar 21 08:39:31 crc kubenswrapper[4696]: I0321 08:39:31.660285 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-rp8tc" Mar 21 08:39:54 crc kubenswrapper[4696]: I0321 08:39:54.547905 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh"] Mar 21 08:39:54 crc kubenswrapper[4696]: I0321 08:39:54.549695 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh" Mar 21 08:39:54 crc kubenswrapper[4696]: I0321 08:39:54.555751 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 21 08:39:54 crc kubenswrapper[4696]: I0321 08:39:54.568309 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh"] Mar 21 08:39:54 crc kubenswrapper[4696]: I0321 08:39:54.617310 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwwcm\" (UniqueName: \"kubernetes.io/projected/8c8ec8ba-d404-4dca-aac7-3619ef7dbd75-kube-api-access-dwwcm\") pod \"b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh\" (UID: \"8c8ec8ba-d404-4dca-aac7-3619ef7dbd75\") " pod="openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh" Mar 21 08:39:54 crc kubenswrapper[4696]: I0321 08:39:54.617362 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8c8ec8ba-d404-4dca-aac7-3619ef7dbd75-bundle\") pod \"b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh\" (UID: \"8c8ec8ba-d404-4dca-aac7-3619ef7dbd75\") " pod="openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh" Mar 21 08:39:54 crc kubenswrapper[4696]: I0321 08:39:54.617524 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8c8ec8ba-d404-4dca-aac7-3619ef7dbd75-util\") pod \"b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh\" (UID: \"8c8ec8ba-d404-4dca-aac7-3619ef7dbd75\") " pod="openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh" Mar 21 08:39:54 crc kubenswrapper[4696]: I0321 08:39:54.719106 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8c8ec8ba-d404-4dca-aac7-3619ef7dbd75-bundle\") pod \"b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh\" (UID: \"8c8ec8ba-d404-4dca-aac7-3619ef7dbd75\") " pod="openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh" Mar 21 08:39:54 crc kubenswrapper[4696]: I0321 08:39:54.719498 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8c8ec8ba-d404-4dca-aac7-3619ef7dbd75-util\") pod \"b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh\" (UID: \"8c8ec8ba-d404-4dca-aac7-3619ef7dbd75\") " pod="openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh" Mar 21 08:39:54 crc kubenswrapper[4696]: I0321 08:39:54.719676 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwwcm\" (UniqueName: \"kubernetes.io/projected/8c8ec8ba-d404-4dca-aac7-3619ef7dbd75-kube-api-access-dwwcm\") pod \"b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh\" (UID: \"8c8ec8ba-d404-4dca-aac7-3619ef7dbd75\") " pod="openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh" Mar 21 08:39:54 crc kubenswrapper[4696]: I0321 08:39:54.719784 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8c8ec8ba-d404-4dca-aac7-3619ef7dbd75-bundle\") pod \"b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh\" (UID: \"8c8ec8ba-d404-4dca-aac7-3619ef7dbd75\") " pod="openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh" Mar 21 08:39:54 crc kubenswrapper[4696]: I0321 08:39:54.719809 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8c8ec8ba-d404-4dca-aac7-3619ef7dbd75-util\") pod \"b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh\" (UID: \"8c8ec8ba-d404-4dca-aac7-3619ef7dbd75\") " pod="openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh" Mar 21 08:39:54 crc kubenswrapper[4696]: I0321 08:39:54.737251 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwwcm\" (UniqueName: \"kubernetes.io/projected/8c8ec8ba-d404-4dca-aac7-3619ef7dbd75-kube-api-access-dwwcm\") pod \"b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh\" (UID: \"8c8ec8ba-d404-4dca-aac7-3619ef7dbd75\") " pod="openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh" Mar 21 08:39:54 crc kubenswrapper[4696]: I0321 08:39:54.868039 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh" Mar 21 08:39:55 crc kubenswrapper[4696]: I0321 08:39:55.343997 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh"] Mar 21 08:39:55 crc kubenswrapper[4696]: I0321 08:39:55.857873 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh" event={"ID":"8c8ec8ba-d404-4dca-aac7-3619ef7dbd75","Type":"ContainerStarted","Data":"b2b85bde376e5ffbd302e15e4399e96925d615bea0fc274fc86aa89b6ee4dd0f"} Mar 21 08:39:55 crc kubenswrapper[4696]: I0321 08:39:55.858216 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh" event={"ID":"8c8ec8ba-d404-4dca-aac7-3619ef7dbd75","Type":"ContainerStarted","Data":"f2e8355514f3dbc730aabf59b5afdbf3653ba7e310925d4e5a1f41578fa4522c"} Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.335390 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.336078 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.338015 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.338249 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.350955 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.478036 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-13b2880e-cf96-4a62-8ad5-0bdaad5f6956\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-13b2880e-cf96-4a62-8ad5-0bdaad5f6956\") pod \"minio\" (UID: \"b167f303-3b1c-4c58-92e0-0cf5beedf392\") " pod="minio-dev/minio" Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.478335 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4nw4\" (UniqueName: \"kubernetes.io/projected/b167f303-3b1c-4c58-92e0-0cf5beedf392-kube-api-access-m4nw4\") pod \"minio\" (UID: \"b167f303-3b1c-4c58-92e0-0cf5beedf392\") " pod="minio-dev/minio" Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.579584 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-13b2880e-cf96-4a62-8ad5-0bdaad5f6956\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-13b2880e-cf96-4a62-8ad5-0bdaad5f6956\") pod \"minio\" (UID: \"b167f303-3b1c-4c58-92e0-0cf5beedf392\") " pod="minio-dev/minio" Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.579680 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4nw4\" (UniqueName: \"kubernetes.io/projected/b167f303-3b1c-4c58-92e0-0cf5beedf392-kube-api-access-m4nw4\") pod \"minio\" (UID: \"b167f303-3b1c-4c58-92e0-0cf5beedf392\") " pod="minio-dev/minio" Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.583631 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.583736 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-13b2880e-cf96-4a62-8ad5-0bdaad5f6956\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-13b2880e-cf96-4a62-8ad5-0bdaad5f6956\") pod \"minio\" (UID: \"b167f303-3b1c-4c58-92e0-0cf5beedf392\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bf10dfaec65f0264732a94e2a43abd85777d66f5d12b374f3f7f85214ab43a4e/globalmount\"" pod="minio-dev/minio" Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.606676 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-13b2880e-cf96-4a62-8ad5-0bdaad5f6956\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-13b2880e-cf96-4a62-8ad5-0bdaad5f6956\") pod \"minio\" (UID: \"b167f303-3b1c-4c58-92e0-0cf5beedf392\") " pod="minio-dev/minio" Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.607608 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4nw4\" (UniqueName: \"kubernetes.io/projected/b167f303-3b1c-4c58-92e0-0cf5beedf392-kube-api-access-m4nw4\") pod \"minio\" (UID: \"b167f303-3b1c-4c58-92e0-0cf5beedf392\") " pod="minio-dev/minio" Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.682367 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.865883 4696 generic.go:334] "Generic (PLEG): container finished" podID="8c8ec8ba-d404-4dca-aac7-3619ef7dbd75" containerID="b2b85bde376e5ffbd302e15e4399e96925d615bea0fc274fc86aa89b6ee4dd0f" exitCode=0 Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.865958 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh" event={"ID":"8c8ec8ba-d404-4dca-aac7-3619ef7dbd75","Type":"ContainerDied","Data":"b2b85bde376e5ffbd302e15e4399e96925d615bea0fc274fc86aa89b6ee4dd0f"} Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.898605 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7wm98"] Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.904465 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7wm98" Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.915778 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7wm98"] Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.988987 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6b2q\" (UniqueName: \"kubernetes.io/projected/3897cce3-d8d6-468a-abc8-2d49d40b2eb8-kube-api-access-m6b2q\") pod \"redhat-operators-7wm98\" (UID: \"3897cce3-d8d6-468a-abc8-2d49d40b2eb8\") " pod="openshift-marketplace/redhat-operators-7wm98" Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.989032 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3897cce3-d8d6-468a-abc8-2d49d40b2eb8-catalog-content\") pod \"redhat-operators-7wm98\" (UID: \"3897cce3-d8d6-468a-abc8-2d49d40b2eb8\") " pod="openshift-marketplace/redhat-operators-7wm98" Mar 21 08:39:56 crc kubenswrapper[4696]: I0321 08:39:56.989056 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3897cce3-d8d6-468a-abc8-2d49d40b2eb8-utilities\") pod \"redhat-operators-7wm98\" (UID: \"3897cce3-d8d6-468a-abc8-2d49d40b2eb8\") " pod="openshift-marketplace/redhat-operators-7wm98" Mar 21 08:39:57 crc kubenswrapper[4696]: I0321 08:39:57.090353 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6b2q\" (UniqueName: \"kubernetes.io/projected/3897cce3-d8d6-468a-abc8-2d49d40b2eb8-kube-api-access-m6b2q\") pod \"redhat-operators-7wm98\" (UID: \"3897cce3-d8d6-468a-abc8-2d49d40b2eb8\") " pod="openshift-marketplace/redhat-operators-7wm98" Mar 21 08:39:57 crc kubenswrapper[4696]: I0321 08:39:57.090397 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3897cce3-d8d6-468a-abc8-2d49d40b2eb8-catalog-content\") pod \"redhat-operators-7wm98\" (UID: \"3897cce3-d8d6-468a-abc8-2d49d40b2eb8\") " pod="openshift-marketplace/redhat-operators-7wm98" Mar 21 08:39:57 crc kubenswrapper[4696]: I0321 08:39:57.090416 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3897cce3-d8d6-468a-abc8-2d49d40b2eb8-utilities\") pod \"redhat-operators-7wm98\" (UID: \"3897cce3-d8d6-468a-abc8-2d49d40b2eb8\") " pod="openshift-marketplace/redhat-operators-7wm98" Mar 21 08:39:57 crc kubenswrapper[4696]: I0321 08:39:57.090934 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3897cce3-d8d6-468a-abc8-2d49d40b2eb8-catalog-content\") pod \"redhat-operators-7wm98\" (UID: \"3897cce3-d8d6-468a-abc8-2d49d40b2eb8\") " pod="openshift-marketplace/redhat-operators-7wm98" Mar 21 08:39:57 crc kubenswrapper[4696]: I0321 08:39:57.090970 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3897cce3-d8d6-468a-abc8-2d49d40b2eb8-utilities\") pod \"redhat-operators-7wm98\" (UID: \"3897cce3-d8d6-468a-abc8-2d49d40b2eb8\") " pod="openshift-marketplace/redhat-operators-7wm98" Mar 21 08:39:57 crc kubenswrapper[4696]: I0321 08:39:57.095472 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Mar 21 08:39:57 crc kubenswrapper[4696]: I0321 08:39:57.111417 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6b2q\" (UniqueName: \"kubernetes.io/projected/3897cce3-d8d6-468a-abc8-2d49d40b2eb8-kube-api-access-m6b2q\") pod \"redhat-operators-7wm98\" (UID: \"3897cce3-d8d6-468a-abc8-2d49d40b2eb8\") " pod="openshift-marketplace/redhat-operators-7wm98" Mar 21 08:39:57 crc kubenswrapper[4696]: I0321 08:39:57.227270 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7wm98" Mar 21 08:39:57 crc kubenswrapper[4696]: I0321 08:39:57.703221 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7wm98"] Mar 21 08:39:57 crc kubenswrapper[4696]: W0321 08:39:57.708466 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3897cce3_d8d6_468a_abc8_2d49d40b2eb8.slice/crio-58ba01dce1b45444904b3618f68de920ac64dd59650489f46d9f64dac5b9321f WatchSource:0}: Error finding container 58ba01dce1b45444904b3618f68de920ac64dd59650489f46d9f64dac5b9321f: Status 404 returned error can't find the container with id 58ba01dce1b45444904b3618f68de920ac64dd59650489f46d9f64dac5b9321f Mar 21 08:39:57 crc kubenswrapper[4696]: I0321 08:39:57.874778 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7wm98" event={"ID":"3897cce3-d8d6-468a-abc8-2d49d40b2eb8","Type":"ContainerStarted","Data":"1e5c0ddd234bc8e87810c45461da996f293cc7c1c84130ddf99f14c1e9bc0c88"} Mar 21 08:39:57 crc kubenswrapper[4696]: I0321 08:39:57.874840 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7wm98" event={"ID":"3897cce3-d8d6-468a-abc8-2d49d40b2eb8","Type":"ContainerStarted","Data":"58ba01dce1b45444904b3618f68de920ac64dd59650489f46d9f64dac5b9321f"} Mar 21 08:39:57 crc kubenswrapper[4696]: I0321 08:39:57.877198 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"b167f303-3b1c-4c58-92e0-0cf5beedf392","Type":"ContainerStarted","Data":"dc30e0213c460ed0b0a88054d65527ae72865dc90debd8d140e68127136e2faf"} Mar 21 08:39:58 crc kubenswrapper[4696]: I0321 08:39:58.886403 4696 generic.go:334] "Generic (PLEG): container finished" podID="8c8ec8ba-d404-4dca-aac7-3619ef7dbd75" containerID="88b9e758581f1f6a73c66e6e588862050aefe6421450c04780efe1a6e0c51ed1" exitCode=0 Mar 21 08:39:58 crc kubenswrapper[4696]: I0321 08:39:58.886877 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh" event={"ID":"8c8ec8ba-d404-4dca-aac7-3619ef7dbd75","Type":"ContainerDied","Data":"88b9e758581f1f6a73c66e6e588862050aefe6421450c04780efe1a6e0c51ed1"} Mar 21 08:39:58 crc kubenswrapper[4696]: I0321 08:39:58.890738 4696 generic.go:334] "Generic (PLEG): container finished" podID="3897cce3-d8d6-468a-abc8-2d49d40b2eb8" containerID="1e5c0ddd234bc8e87810c45461da996f293cc7c1c84130ddf99f14c1e9bc0c88" exitCode=0 Mar 21 08:39:58 crc kubenswrapper[4696]: I0321 08:39:58.890796 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7wm98" event={"ID":"3897cce3-d8d6-468a-abc8-2d49d40b2eb8","Type":"ContainerDied","Data":"1e5c0ddd234bc8e87810c45461da996f293cc7c1c84130ddf99f14c1e9bc0c88"} Mar 21 08:39:59 crc kubenswrapper[4696]: I0321 08:39:59.897470 4696 generic.go:334] "Generic (PLEG): container finished" podID="8c8ec8ba-d404-4dca-aac7-3619ef7dbd75" containerID="4f799167cb1cbac03ff202c53458cd60d431fc4715e93f6e9ef01a736e5317ad" exitCode=0 Mar 21 08:39:59 crc kubenswrapper[4696]: I0321 08:39:59.897507 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh" event={"ID":"8c8ec8ba-d404-4dca-aac7-3619ef7dbd75","Type":"ContainerDied","Data":"4f799167cb1cbac03ff202c53458cd60d431fc4715e93f6e9ef01a736e5317ad"} Mar 21 08:40:00 crc kubenswrapper[4696]: I0321 08:40:00.128961 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568040-n66dl"] Mar 21 08:40:00 crc kubenswrapper[4696]: I0321 08:40:00.129755 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568040-n66dl" Mar 21 08:40:00 crc kubenswrapper[4696]: I0321 08:40:00.132869 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 08:40:00 crc kubenswrapper[4696]: I0321 08:40:00.135994 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 08:40:00 crc kubenswrapper[4696]: I0321 08:40:00.136158 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 08:40:00 crc kubenswrapper[4696]: I0321 08:40:00.138879 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568040-n66dl"] Mar 21 08:40:00 crc kubenswrapper[4696]: I0321 08:40:00.183981 4696 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 21 08:40:00 crc kubenswrapper[4696]: I0321 08:40:00.243729 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2h7c\" (UniqueName: \"kubernetes.io/projected/e1d6d6a5-6b27-4b94-b754-2544992c5622-kube-api-access-q2h7c\") pod \"auto-csr-approver-29568040-n66dl\" (UID: \"e1d6d6a5-6b27-4b94-b754-2544992c5622\") " pod="openshift-infra/auto-csr-approver-29568040-n66dl" Mar 21 08:40:00 crc kubenswrapper[4696]: I0321 08:40:00.345697 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2h7c\" (UniqueName: \"kubernetes.io/projected/e1d6d6a5-6b27-4b94-b754-2544992c5622-kube-api-access-q2h7c\") pod \"auto-csr-approver-29568040-n66dl\" (UID: \"e1d6d6a5-6b27-4b94-b754-2544992c5622\") " pod="openshift-infra/auto-csr-approver-29568040-n66dl" Mar 21 08:40:00 crc kubenswrapper[4696]: I0321 08:40:00.377860 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2h7c\" (UniqueName: \"kubernetes.io/projected/e1d6d6a5-6b27-4b94-b754-2544992c5622-kube-api-access-q2h7c\") pod \"auto-csr-approver-29568040-n66dl\" (UID: \"e1d6d6a5-6b27-4b94-b754-2544992c5622\") " pod="openshift-infra/auto-csr-approver-29568040-n66dl" Mar 21 08:40:00 crc kubenswrapper[4696]: I0321 08:40:00.476665 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568040-n66dl" Mar 21 08:40:01 crc kubenswrapper[4696]: I0321 08:40:01.129212 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh" Mar 21 08:40:01 crc kubenswrapper[4696]: I0321 08:40:01.260649 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwwcm\" (UniqueName: \"kubernetes.io/projected/8c8ec8ba-d404-4dca-aac7-3619ef7dbd75-kube-api-access-dwwcm\") pod \"8c8ec8ba-d404-4dca-aac7-3619ef7dbd75\" (UID: \"8c8ec8ba-d404-4dca-aac7-3619ef7dbd75\") " Mar 21 08:40:01 crc kubenswrapper[4696]: I0321 08:40:01.260789 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8c8ec8ba-d404-4dca-aac7-3619ef7dbd75-util\") pod \"8c8ec8ba-d404-4dca-aac7-3619ef7dbd75\" (UID: \"8c8ec8ba-d404-4dca-aac7-3619ef7dbd75\") " Mar 21 08:40:01 crc kubenswrapper[4696]: I0321 08:40:01.260838 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8c8ec8ba-d404-4dca-aac7-3619ef7dbd75-bundle\") pod \"8c8ec8ba-d404-4dca-aac7-3619ef7dbd75\" (UID: \"8c8ec8ba-d404-4dca-aac7-3619ef7dbd75\") " Mar 21 08:40:01 crc kubenswrapper[4696]: I0321 08:40:01.262652 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c8ec8ba-d404-4dca-aac7-3619ef7dbd75-bundle" (OuterVolumeSpecName: "bundle") pod "8c8ec8ba-d404-4dca-aac7-3619ef7dbd75" (UID: "8c8ec8ba-d404-4dca-aac7-3619ef7dbd75"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:40:01 crc kubenswrapper[4696]: I0321 08:40:01.267697 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c8ec8ba-d404-4dca-aac7-3619ef7dbd75-kube-api-access-dwwcm" (OuterVolumeSpecName: "kube-api-access-dwwcm") pod "8c8ec8ba-d404-4dca-aac7-3619ef7dbd75" (UID: "8c8ec8ba-d404-4dca-aac7-3619ef7dbd75"). InnerVolumeSpecName "kube-api-access-dwwcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:40:01 crc kubenswrapper[4696]: I0321 08:40:01.274946 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c8ec8ba-d404-4dca-aac7-3619ef7dbd75-util" (OuterVolumeSpecName: "util") pod "8c8ec8ba-d404-4dca-aac7-3619ef7dbd75" (UID: "8c8ec8ba-d404-4dca-aac7-3619ef7dbd75"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:40:01 crc kubenswrapper[4696]: I0321 08:40:01.361937 4696 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8c8ec8ba-d404-4dca-aac7-3619ef7dbd75-util\") on node \"crc\" DevicePath \"\"" Mar 21 08:40:01 crc kubenswrapper[4696]: I0321 08:40:01.361975 4696 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8c8ec8ba-d404-4dca-aac7-3619ef7dbd75-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:40:01 crc kubenswrapper[4696]: I0321 08:40:01.361988 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwwcm\" (UniqueName: \"kubernetes.io/projected/8c8ec8ba-d404-4dca-aac7-3619ef7dbd75-kube-api-access-dwwcm\") on node \"crc\" DevicePath \"\"" Mar 21 08:40:01 crc kubenswrapper[4696]: I0321 08:40:01.491375 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568040-n66dl"] Mar 21 08:40:01 crc kubenswrapper[4696]: I0321 08:40:01.910048 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568040-n66dl" event={"ID":"e1d6d6a5-6b27-4b94-b754-2544992c5622","Type":"ContainerStarted","Data":"8c671b08f8952011f5f37ec552c29a68756d07be4328dd9325c51feb5be21920"} Mar 21 08:40:01 crc kubenswrapper[4696]: I0321 08:40:01.913323 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh" event={"ID":"8c8ec8ba-d404-4dca-aac7-3619ef7dbd75","Type":"ContainerDied","Data":"f2e8355514f3dbc730aabf59b5afdbf3653ba7e310925d4e5a1f41578fa4522c"} Mar 21 08:40:01 crc kubenswrapper[4696]: I0321 08:40:01.913394 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2e8355514f3dbc730aabf59b5afdbf3653ba7e310925d4e5a1f41578fa4522c" Mar 21 08:40:01 crc kubenswrapper[4696]: I0321 08:40:01.913412 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh" Mar 21 08:40:01 crc kubenswrapper[4696]: I0321 08:40:01.915312 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"b167f303-3b1c-4c58-92e0-0cf5beedf392","Type":"ContainerStarted","Data":"183e2ea020d3d90084208f1f4bdf482c90ec5ba58bc7ee73a6b7908b85cf3fdc"} Mar 21 08:40:01 crc kubenswrapper[4696]: I0321 08:40:01.919105 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7wm98" event={"ID":"3897cce3-d8d6-468a-abc8-2d49d40b2eb8","Type":"ContainerStarted","Data":"ea0c1d84f69f5a7f14da8706bbfd0d9056852f4fb96f92628e5371b9eb2e0557"} Mar 21 08:40:01 crc kubenswrapper[4696]: I0321 08:40:01.950843 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=4.920744398 podStartE2EDuration="8.950810073s" podCreationTimestamp="2026-03-21 08:39:53 +0000 UTC" firstStartedPulling="2026-03-21 08:39:57.106011066 +0000 UTC m=+731.226891779" lastFinishedPulling="2026-03-21 08:40:01.136076741 +0000 UTC m=+735.256957454" observedRunningTime="2026-03-21 08:40:01.942307383 +0000 UTC m=+736.063188096" watchObservedRunningTime="2026-03-21 08:40:01.950810073 +0000 UTC m=+736.071690786" Mar 21 08:40:02 crc kubenswrapper[4696]: I0321 08:40:02.927014 4696 generic.go:334] "Generic (PLEG): container finished" podID="3897cce3-d8d6-468a-abc8-2d49d40b2eb8" containerID="ea0c1d84f69f5a7f14da8706bbfd0d9056852f4fb96f92628e5371b9eb2e0557" exitCode=0 Mar 21 08:40:02 crc kubenswrapper[4696]: I0321 08:40:02.927063 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7wm98" event={"ID":"3897cce3-d8d6-468a-abc8-2d49d40b2eb8","Type":"ContainerDied","Data":"ea0c1d84f69f5a7f14da8706bbfd0d9056852f4fb96f92628e5371b9eb2e0557"} Mar 21 08:40:02 crc kubenswrapper[4696]: I0321 08:40:02.929783 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568040-n66dl" event={"ID":"e1d6d6a5-6b27-4b94-b754-2544992c5622","Type":"ContainerStarted","Data":"94325b9c899893b0d183be3caa5acc3b05b5cdcc90d52a9b47f06593563370b2"} Mar 21 08:40:02 crc kubenswrapper[4696]: I0321 08:40:02.981557 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29568040-n66dl" podStartSLOduration=2.065681206 podStartE2EDuration="2.981538597s" podCreationTimestamp="2026-03-21 08:40:00 +0000 UTC" firstStartedPulling="2026-03-21 08:40:01.50151414 +0000 UTC m=+735.622394853" lastFinishedPulling="2026-03-21 08:40:02.417371531 +0000 UTC m=+736.538252244" observedRunningTime="2026-03-21 08:40:02.97741345 +0000 UTC m=+737.098294173" watchObservedRunningTime="2026-03-21 08:40:02.981538597 +0000 UTC m=+737.102419310" Mar 21 08:40:03 crc kubenswrapper[4696]: I0321 08:40:03.936291 4696 generic.go:334] "Generic (PLEG): container finished" podID="e1d6d6a5-6b27-4b94-b754-2544992c5622" containerID="94325b9c899893b0d183be3caa5acc3b05b5cdcc90d52a9b47f06593563370b2" exitCode=0 Mar 21 08:40:03 crc kubenswrapper[4696]: I0321 08:40:03.936335 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568040-n66dl" event={"ID":"e1d6d6a5-6b27-4b94-b754-2544992c5622","Type":"ContainerDied","Data":"94325b9c899893b0d183be3caa5acc3b05b5cdcc90d52a9b47f06593563370b2"} Mar 21 08:40:03 crc kubenswrapper[4696]: I0321 08:40:03.939204 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7wm98" event={"ID":"3897cce3-d8d6-468a-abc8-2d49d40b2eb8","Type":"ContainerStarted","Data":"e5d5478b8c5684b07518cdebf53709b8d0bb8708c9c0f401212395445329397a"} Mar 21 08:40:03 crc kubenswrapper[4696]: I0321 08:40:03.979546 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7wm98" podStartSLOduration=3.408862137 podStartE2EDuration="7.979527712s" podCreationTimestamp="2026-03-21 08:39:56 +0000 UTC" firstStartedPulling="2026-03-21 08:39:58.92329187 +0000 UTC m=+733.044172593" lastFinishedPulling="2026-03-21 08:40:03.493957435 +0000 UTC m=+737.614838168" observedRunningTime="2026-03-21 08:40:03.976746526 +0000 UTC m=+738.097627239" watchObservedRunningTime="2026-03-21 08:40:03.979527712 +0000 UTC m=+738.100408425" Mar 21 08:40:05 crc kubenswrapper[4696]: I0321 08:40:05.168762 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568040-n66dl" Mar 21 08:40:05 crc kubenswrapper[4696]: I0321 08:40:05.313400 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2h7c\" (UniqueName: \"kubernetes.io/projected/e1d6d6a5-6b27-4b94-b754-2544992c5622-kube-api-access-q2h7c\") pod \"e1d6d6a5-6b27-4b94-b754-2544992c5622\" (UID: \"e1d6d6a5-6b27-4b94-b754-2544992c5622\") " Mar 21 08:40:05 crc kubenswrapper[4696]: I0321 08:40:05.332986 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1d6d6a5-6b27-4b94-b754-2544992c5622-kube-api-access-q2h7c" (OuterVolumeSpecName: "kube-api-access-q2h7c") pod "e1d6d6a5-6b27-4b94-b754-2544992c5622" (UID: "e1d6d6a5-6b27-4b94-b754-2544992c5622"). InnerVolumeSpecName "kube-api-access-q2h7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:40:05 crc kubenswrapper[4696]: I0321 08:40:05.414548 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2h7c\" (UniqueName: \"kubernetes.io/projected/e1d6d6a5-6b27-4b94-b754-2544992c5622-kube-api-access-q2h7c\") on node \"crc\" DevicePath \"\"" Mar 21 08:40:05 crc kubenswrapper[4696]: I0321 08:40:05.950897 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568040-n66dl" event={"ID":"e1d6d6a5-6b27-4b94-b754-2544992c5622","Type":"ContainerDied","Data":"8c671b08f8952011f5f37ec552c29a68756d07be4328dd9325c51feb5be21920"} Mar 21 08:40:05 crc kubenswrapper[4696]: I0321 08:40:05.951196 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c671b08f8952011f5f37ec552c29a68756d07be4328dd9325c51feb5be21920" Mar 21 08:40:05 crc kubenswrapper[4696]: I0321 08:40:05.950955 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568040-n66dl" Mar 21 08:40:06 crc kubenswrapper[4696]: I0321 08:40:06.230968 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568034-hmpdb"] Mar 21 08:40:06 crc kubenswrapper[4696]: I0321 08:40:06.234614 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568034-hmpdb"] Mar 21 08:40:06 crc kubenswrapper[4696]: I0321 08:40:06.547032 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fda0225-3507-4265-aae2-8e0e2f8c11d9" path="/var/lib/kubelet/pods/6fda0225-3507-4265-aae2-8e0e2f8c11d9/volumes" Mar 21 08:40:07 crc kubenswrapper[4696]: I0321 08:40:07.228129 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7wm98" Mar 21 08:40:07 crc kubenswrapper[4696]: I0321 08:40:07.229671 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7wm98" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.077132 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx"] Mar 21 08:40:08 crc kubenswrapper[4696]: E0321 08:40:08.077416 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c8ec8ba-d404-4dca-aac7-3619ef7dbd75" containerName="extract" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.077431 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c8ec8ba-d404-4dca-aac7-3619ef7dbd75" containerName="extract" Mar 21 08:40:08 crc kubenswrapper[4696]: E0321 08:40:08.077453 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c8ec8ba-d404-4dca-aac7-3619ef7dbd75" containerName="util" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.077461 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c8ec8ba-d404-4dca-aac7-3619ef7dbd75" containerName="util" Mar 21 08:40:08 crc kubenswrapper[4696]: E0321 08:40:08.077473 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c8ec8ba-d404-4dca-aac7-3619ef7dbd75" containerName="pull" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.077480 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c8ec8ba-d404-4dca-aac7-3619ef7dbd75" containerName="pull" Mar 21 08:40:08 crc kubenswrapper[4696]: E0321 08:40:08.077499 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1d6d6a5-6b27-4b94-b754-2544992c5622" containerName="oc" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.077506 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1d6d6a5-6b27-4b94-b754-2544992c5622" containerName="oc" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.077616 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c8ec8ba-d404-4dca-aac7-3619ef7dbd75" containerName="extract" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.077629 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1d6d6a5-6b27-4b94-b754-2544992c5622" containerName="oc" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.078471 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" Mar 21 08:40:08 crc kubenswrapper[4696]: W0321 08:40:08.081981 4696 reflector.go:561] object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-8nkrs": failed to list *v1.Secret: secrets "loki-operator-controller-manager-dockercfg-8nkrs" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-operators-redhat": no relationship found between node 'crc' and this object Mar 21 08:40:08 crc kubenswrapper[4696]: E0321 08:40:08.082029 4696 reflector.go:158] "Unhandled Error" err="object-\"openshift-operators-redhat\"/\"loki-operator-controller-manager-dockercfg-8nkrs\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"loki-operator-controller-manager-dockercfg-8nkrs\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-operators-redhat\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 21 08:40:08 crc kubenswrapper[4696]: W0321 08:40:08.081991 4696 reflector.go:561] object-"openshift-operators-redhat"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-operators-redhat": no relationship found between node 'crc' and this object Mar 21 08:40:08 crc kubenswrapper[4696]: E0321 08:40:08.082064 4696 reflector.go:158] "Unhandled Error" err="object-\"openshift-operators-redhat\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-operators-redhat\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 21 08:40:08 crc kubenswrapper[4696]: W0321 08:40:08.082228 4696 reflector.go:561] object-"openshift-operators-redhat"/"loki-operator-manager-config": failed to list *v1.ConfigMap: configmaps "loki-operator-manager-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-operators-redhat": no relationship found between node 'crc' and this object Mar 21 08:40:08 crc kubenswrapper[4696]: E0321 08:40:08.082256 4696 reflector.go:158] "Unhandled Error" err="object-\"openshift-operators-redhat\"/\"loki-operator-manager-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"loki-operator-manager-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-operators-redhat\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 21 08:40:08 crc kubenswrapper[4696]: W0321 08:40:08.082353 4696 reflector.go:561] object-"openshift-operators-redhat"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-operators-redhat": no relationship found between node 'crc' and this object Mar 21 08:40:08 crc kubenswrapper[4696]: E0321 08:40:08.082380 4696 reflector.go:158] "Unhandled Error" err="object-\"openshift-operators-redhat\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-operators-redhat\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.082391 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Mar 21 08:40:08 crc kubenswrapper[4696]: W0321 08:40:08.084120 4696 reflector.go:561] object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert": failed to list *v1.Secret: secrets "loki-operator-controller-manager-service-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-operators-redhat": no relationship found between node 'crc' and this object Mar 21 08:40:08 crc kubenswrapper[4696]: E0321 08:40:08.084153 4696 reflector.go:158] "Unhandled Error" err="object-\"openshift-operators-redhat\"/\"loki-operator-controller-manager-service-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"loki-operator-controller-manager-service-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-operators-redhat\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.102166 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx"] Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.145901 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/75d1fdab-70ac-46df-8b60-3e2f75985686-manager-config\") pod \"loki-operator-controller-manager-5685547f78-5x2wx\" (UID: \"75d1fdab-70ac-46df-8b60-3e2f75985686\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.145953 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/75d1fdab-70ac-46df-8b60-3e2f75985686-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-5685547f78-5x2wx\" (UID: \"75d1fdab-70ac-46df-8b60-3e2f75985686\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.146113 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztj5r\" (UniqueName: \"kubernetes.io/projected/75d1fdab-70ac-46df-8b60-3e2f75985686-kube-api-access-ztj5r\") pod \"loki-operator-controller-manager-5685547f78-5x2wx\" (UID: \"75d1fdab-70ac-46df-8b60-3e2f75985686\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.146176 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/75d1fdab-70ac-46df-8b60-3e2f75985686-apiservice-cert\") pod \"loki-operator-controller-manager-5685547f78-5x2wx\" (UID: \"75d1fdab-70ac-46df-8b60-3e2f75985686\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.146239 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/75d1fdab-70ac-46df-8b60-3e2f75985686-webhook-cert\") pod \"loki-operator-controller-manager-5685547f78-5x2wx\" (UID: \"75d1fdab-70ac-46df-8b60-3e2f75985686\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.247870 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztj5r\" (UniqueName: \"kubernetes.io/projected/75d1fdab-70ac-46df-8b60-3e2f75985686-kube-api-access-ztj5r\") pod \"loki-operator-controller-manager-5685547f78-5x2wx\" (UID: \"75d1fdab-70ac-46df-8b60-3e2f75985686\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.247946 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/75d1fdab-70ac-46df-8b60-3e2f75985686-apiservice-cert\") pod \"loki-operator-controller-manager-5685547f78-5x2wx\" (UID: \"75d1fdab-70ac-46df-8b60-3e2f75985686\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.247991 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/75d1fdab-70ac-46df-8b60-3e2f75985686-webhook-cert\") pod \"loki-operator-controller-manager-5685547f78-5x2wx\" (UID: \"75d1fdab-70ac-46df-8b60-3e2f75985686\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.248053 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/75d1fdab-70ac-46df-8b60-3e2f75985686-manager-config\") pod \"loki-operator-controller-manager-5685547f78-5x2wx\" (UID: \"75d1fdab-70ac-46df-8b60-3e2f75985686\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.248080 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/75d1fdab-70ac-46df-8b60-3e2f75985686-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-5685547f78-5x2wx\" (UID: \"75d1fdab-70ac-46df-8b60-3e2f75985686\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.259949 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/75d1fdab-70ac-46df-8b60-3e2f75985686-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-5685547f78-5x2wx\" (UID: \"75d1fdab-70ac-46df-8b60-3e2f75985686\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.279846 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7wm98" podUID="3897cce3-d8d6-468a-abc8-2d49d40b2eb8" containerName="registry-server" probeResult="failure" output=< Mar 21 08:40:08 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Mar 21 08:40:08 crc kubenswrapper[4696]: > Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.907760 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.913366 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/75d1fdab-70ac-46df-8b60-3e2f75985686-apiservice-cert\") pod \"loki-operator-controller-manager-5685547f78-5x2wx\" (UID: \"75d1fdab-70ac-46df-8b60-3e2f75985686\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.913493 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/75d1fdab-70ac-46df-8b60-3e2f75985686-webhook-cert\") pod \"loki-operator-controller-manager-5685547f78-5x2wx\" (UID: \"75d1fdab-70ac-46df-8b60-3e2f75985686\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" Mar 21 08:40:08 crc kubenswrapper[4696]: I0321 08:40:08.981773 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Mar 21 08:40:09 crc kubenswrapper[4696]: I0321 08:40:09.008536 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-8nkrs" Mar 21 08:40:09 crc kubenswrapper[4696]: E0321 08:40:09.248906 4696 configmap.go:193] Couldn't get configMap openshift-operators-redhat/loki-operator-manager-config: failed to sync configmap cache: timed out waiting for the condition Mar 21 08:40:09 crc kubenswrapper[4696]: E0321 08:40:09.249020 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/75d1fdab-70ac-46df-8b60-3e2f75985686-manager-config podName:75d1fdab-70ac-46df-8b60-3e2f75985686 nodeName:}" failed. No retries permitted until 2026-03-21 08:40:09.74899423 +0000 UTC m=+743.869874943 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "manager-config" (UniqueName: "kubernetes.io/configmap/75d1fdab-70ac-46df-8b60-3e2f75985686-manager-config") pod "loki-operator-controller-manager-5685547f78-5x2wx" (UID: "75d1fdab-70ac-46df-8b60-3e2f75985686") : failed to sync configmap cache: timed out waiting for the condition Mar 21 08:40:09 crc kubenswrapper[4696]: E0321 08:40:09.274008 4696 projected.go:288] Couldn't get configMap openshift-operators-redhat/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 21 08:40:09 crc kubenswrapper[4696]: E0321 08:40:09.274063 4696 projected.go:194] Error preparing data for projected volume kube-api-access-ztj5r for pod openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx: failed to sync configmap cache: timed out waiting for the condition Mar 21 08:40:09 crc kubenswrapper[4696]: E0321 08:40:09.274123 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/75d1fdab-70ac-46df-8b60-3e2f75985686-kube-api-access-ztj5r podName:75d1fdab-70ac-46df-8b60-3e2f75985686 nodeName:}" failed. No retries permitted until 2026-03-21 08:40:09.774106481 +0000 UTC m=+743.894987194 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-ztj5r" (UniqueName: "kubernetes.io/projected/75d1fdab-70ac-46df-8b60-3e2f75985686-kube-api-access-ztj5r") pod "loki-operator-controller-manager-5685547f78-5x2wx" (UID: "75d1fdab-70ac-46df-8b60-3e2f75985686") : failed to sync configmap cache: timed out waiting for the condition Mar 21 08:40:09 crc kubenswrapper[4696]: I0321 08:40:09.539025 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Mar 21 08:40:09 crc kubenswrapper[4696]: I0321 08:40:09.677891 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Mar 21 08:40:09 crc kubenswrapper[4696]: I0321 08:40:09.765190 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/75d1fdab-70ac-46df-8b60-3e2f75985686-manager-config\") pod \"loki-operator-controller-manager-5685547f78-5x2wx\" (UID: \"75d1fdab-70ac-46df-8b60-3e2f75985686\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" Mar 21 08:40:09 crc kubenswrapper[4696]: I0321 08:40:09.766094 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/75d1fdab-70ac-46df-8b60-3e2f75985686-manager-config\") pod \"loki-operator-controller-manager-5685547f78-5x2wx\" (UID: \"75d1fdab-70ac-46df-8b60-3e2f75985686\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" Mar 21 08:40:09 crc kubenswrapper[4696]: I0321 08:40:09.866067 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztj5r\" (UniqueName: \"kubernetes.io/projected/75d1fdab-70ac-46df-8b60-3e2f75985686-kube-api-access-ztj5r\") pod \"loki-operator-controller-manager-5685547f78-5x2wx\" (UID: \"75d1fdab-70ac-46df-8b60-3e2f75985686\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" Mar 21 08:40:09 crc kubenswrapper[4696]: I0321 08:40:09.870100 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztj5r\" (UniqueName: \"kubernetes.io/projected/75d1fdab-70ac-46df-8b60-3e2f75985686-kube-api-access-ztj5r\") pod \"loki-operator-controller-manager-5685547f78-5x2wx\" (UID: \"75d1fdab-70ac-46df-8b60-3e2f75985686\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" Mar 21 08:40:09 crc kubenswrapper[4696]: I0321 08:40:09.896068 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" Mar 21 08:40:10 crc kubenswrapper[4696]: I0321 08:40:10.131695 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx"] Mar 21 08:40:10 crc kubenswrapper[4696]: W0321 08:40:10.131837 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75d1fdab_70ac_46df_8b60_3e2f75985686.slice/crio-de7b2d1f249121c9b7ce0753c89799fec53e3f2c02df7371479a872e1bc1031d WatchSource:0}: Error finding container de7b2d1f249121c9b7ce0753c89799fec53e3f2c02df7371479a872e1bc1031d: Status 404 returned error can't find the container with id de7b2d1f249121c9b7ce0753c89799fec53e3f2c02df7371479a872e1bc1031d Mar 21 08:40:10 crc kubenswrapper[4696]: I0321 08:40:10.983162 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" event={"ID":"75d1fdab-70ac-46df-8b60-3e2f75985686","Type":"ContainerStarted","Data":"de7b2d1f249121c9b7ce0753c89799fec53e3f2c02df7371479a872e1bc1031d"} Mar 21 08:40:15 crc kubenswrapper[4696]: I0321 08:40:15.008090 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" event={"ID":"75d1fdab-70ac-46df-8b60-3e2f75985686","Type":"ContainerStarted","Data":"e99c73455d5873774a9d4348e6e27bb65401e6527c74cad181af34489c02a6d4"} Mar 21 08:40:17 crc kubenswrapper[4696]: I0321 08:40:17.279347 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7wm98" Mar 21 08:40:17 crc kubenswrapper[4696]: I0321 08:40:17.324388 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7wm98" Mar 21 08:40:19 crc kubenswrapper[4696]: I0321 08:40:19.489499 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7wm98"] Mar 21 08:40:19 crc kubenswrapper[4696]: I0321 08:40:19.489737 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7wm98" podUID="3897cce3-d8d6-468a-abc8-2d49d40b2eb8" containerName="registry-server" containerID="cri-o://e5d5478b8c5684b07518cdebf53709b8d0bb8708c9c0f401212395445329397a" gracePeriod=2 Mar 21 08:40:20 crc kubenswrapper[4696]: I0321 08:40:20.052072 4696 generic.go:334] "Generic (PLEG): container finished" podID="3897cce3-d8d6-468a-abc8-2d49d40b2eb8" containerID="e5d5478b8c5684b07518cdebf53709b8d0bb8708c9c0f401212395445329397a" exitCode=0 Mar 21 08:40:20 crc kubenswrapper[4696]: I0321 08:40:20.052235 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7wm98" event={"ID":"3897cce3-d8d6-468a-abc8-2d49d40b2eb8","Type":"ContainerDied","Data":"e5d5478b8c5684b07518cdebf53709b8d0bb8708c9c0f401212395445329397a"} Mar 21 08:40:20 crc kubenswrapper[4696]: I0321 08:40:20.368999 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7wm98" Mar 21 08:40:20 crc kubenswrapper[4696]: I0321 08:40:20.518501 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3897cce3-d8d6-468a-abc8-2d49d40b2eb8-utilities\") pod \"3897cce3-d8d6-468a-abc8-2d49d40b2eb8\" (UID: \"3897cce3-d8d6-468a-abc8-2d49d40b2eb8\") " Mar 21 08:40:20 crc kubenswrapper[4696]: I0321 08:40:20.518794 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6b2q\" (UniqueName: \"kubernetes.io/projected/3897cce3-d8d6-468a-abc8-2d49d40b2eb8-kube-api-access-m6b2q\") pod \"3897cce3-d8d6-468a-abc8-2d49d40b2eb8\" (UID: \"3897cce3-d8d6-468a-abc8-2d49d40b2eb8\") " Mar 21 08:40:20 crc kubenswrapper[4696]: I0321 08:40:20.518882 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3897cce3-d8d6-468a-abc8-2d49d40b2eb8-catalog-content\") pod \"3897cce3-d8d6-468a-abc8-2d49d40b2eb8\" (UID: \"3897cce3-d8d6-468a-abc8-2d49d40b2eb8\") " Mar 21 08:40:20 crc kubenswrapper[4696]: I0321 08:40:20.519803 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3897cce3-d8d6-468a-abc8-2d49d40b2eb8-utilities" (OuterVolumeSpecName: "utilities") pod "3897cce3-d8d6-468a-abc8-2d49d40b2eb8" (UID: "3897cce3-d8d6-468a-abc8-2d49d40b2eb8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:40:20 crc kubenswrapper[4696]: I0321 08:40:20.529151 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3897cce3-d8d6-468a-abc8-2d49d40b2eb8-kube-api-access-m6b2q" (OuterVolumeSpecName: "kube-api-access-m6b2q") pod "3897cce3-d8d6-468a-abc8-2d49d40b2eb8" (UID: "3897cce3-d8d6-468a-abc8-2d49d40b2eb8"). InnerVolumeSpecName "kube-api-access-m6b2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:40:20 crc kubenswrapper[4696]: I0321 08:40:20.620515 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3897cce3-d8d6-468a-abc8-2d49d40b2eb8-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 08:40:20 crc kubenswrapper[4696]: I0321 08:40:20.620687 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6b2q\" (UniqueName: \"kubernetes.io/projected/3897cce3-d8d6-468a-abc8-2d49d40b2eb8-kube-api-access-m6b2q\") on node \"crc\" DevicePath \"\"" Mar 21 08:40:20 crc kubenswrapper[4696]: I0321 08:40:20.647064 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3897cce3-d8d6-468a-abc8-2d49d40b2eb8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3897cce3-d8d6-468a-abc8-2d49d40b2eb8" (UID: "3897cce3-d8d6-468a-abc8-2d49d40b2eb8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:40:20 crc kubenswrapper[4696]: I0321 08:40:20.721913 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3897cce3-d8d6-468a-abc8-2d49d40b2eb8-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 08:40:21 crc kubenswrapper[4696]: I0321 08:40:21.058448 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7wm98" event={"ID":"3897cce3-d8d6-468a-abc8-2d49d40b2eb8","Type":"ContainerDied","Data":"58ba01dce1b45444904b3618f68de920ac64dd59650489f46d9f64dac5b9321f"} Mar 21 08:40:21 crc kubenswrapper[4696]: I0321 08:40:21.058485 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7wm98" Mar 21 08:40:21 crc kubenswrapper[4696]: I0321 08:40:21.058491 4696 scope.go:117] "RemoveContainer" containerID="e5d5478b8c5684b07518cdebf53709b8d0bb8708c9c0f401212395445329397a" Mar 21 08:40:21 crc kubenswrapper[4696]: I0321 08:40:21.061264 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" event={"ID":"75d1fdab-70ac-46df-8b60-3e2f75985686","Type":"ContainerStarted","Data":"3725fe8bad8fb402437e6f3b19e432324c10113e5abbbd5e914fad6cbb8c6450"} Mar 21 08:40:21 crc kubenswrapper[4696]: I0321 08:40:21.061838 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" Mar 21 08:40:21 crc kubenswrapper[4696]: I0321 08:40:21.065684 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" Mar 21 08:40:21 crc kubenswrapper[4696]: I0321 08:40:21.076648 4696 scope.go:117] "RemoveContainer" containerID="ea0c1d84f69f5a7f14da8706bbfd0d9056852f4fb96f92628e5371b9eb2e0557" Mar 21 08:40:21 crc kubenswrapper[4696]: I0321 08:40:21.112708 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-5685547f78-5x2wx" podStartSLOduration=2.835228755 podStartE2EDuration="13.11268546s" podCreationTimestamp="2026-03-21 08:40:08 +0000 UTC" firstStartedPulling="2026-03-21 08:40:10.134439065 +0000 UTC m=+744.255319778" lastFinishedPulling="2026-03-21 08:40:20.41189577 +0000 UTC m=+754.532776483" observedRunningTime="2026-03-21 08:40:21.098539037 +0000 UTC m=+755.219419750" watchObservedRunningTime="2026-03-21 08:40:21.11268546 +0000 UTC m=+755.233566173" Mar 21 08:40:21 crc kubenswrapper[4696]: I0321 08:40:21.118682 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7wm98"] Mar 21 08:40:21 crc kubenswrapper[4696]: I0321 08:40:21.121385 4696 scope.go:117] "RemoveContainer" containerID="1e5c0ddd234bc8e87810c45461da996f293cc7c1c84130ddf99f14c1e9bc0c88" Mar 21 08:40:21 crc kubenswrapper[4696]: I0321 08:40:21.122873 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7wm98"] Mar 21 08:40:22 crc kubenswrapper[4696]: I0321 08:40:22.543123 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3897cce3-d8d6-468a-abc8-2d49d40b2eb8" path="/var/lib/kubelet/pods/3897cce3-d8d6-468a-abc8-2d49d40b2eb8/volumes" Mar 21 08:40:30 crc kubenswrapper[4696]: I0321 08:40:30.341316 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:40:30 crc kubenswrapper[4696]: I0321 08:40:30.342054 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:40:47 crc kubenswrapper[4696]: I0321 08:40:47.072436 4696 scope.go:117] "RemoveContainer" containerID="c84e974a0e7a698bfe0737e0d5841ca77b33de4f4faeefc53886d83dc6e5eea7" Mar 21 08:40:49 crc kubenswrapper[4696]: I0321 08:40:49.865411 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8"] Mar 21 08:40:49 crc kubenswrapper[4696]: E0321 08:40:49.866363 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3897cce3-d8d6-468a-abc8-2d49d40b2eb8" containerName="extract-utilities" Mar 21 08:40:49 crc kubenswrapper[4696]: I0321 08:40:49.866430 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="3897cce3-d8d6-468a-abc8-2d49d40b2eb8" containerName="extract-utilities" Mar 21 08:40:49 crc kubenswrapper[4696]: E0321 08:40:49.866489 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3897cce3-d8d6-468a-abc8-2d49d40b2eb8" containerName="registry-server" Mar 21 08:40:49 crc kubenswrapper[4696]: I0321 08:40:49.866540 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="3897cce3-d8d6-468a-abc8-2d49d40b2eb8" containerName="registry-server" Mar 21 08:40:49 crc kubenswrapper[4696]: E0321 08:40:49.866598 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3897cce3-d8d6-468a-abc8-2d49d40b2eb8" containerName="extract-content" Mar 21 08:40:49 crc kubenswrapper[4696]: I0321 08:40:49.866651 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="3897cce3-d8d6-468a-abc8-2d49d40b2eb8" containerName="extract-content" Mar 21 08:40:49 crc kubenswrapper[4696]: I0321 08:40:49.866800 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="3897cce3-d8d6-468a-abc8-2d49d40b2eb8" containerName="registry-server" Mar 21 08:40:49 crc kubenswrapper[4696]: I0321 08:40:49.867564 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8" Mar 21 08:40:49 crc kubenswrapper[4696]: I0321 08:40:49.872169 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 21 08:40:49 crc kubenswrapper[4696]: I0321 08:40:49.879718 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8"] Mar 21 08:40:49 crc kubenswrapper[4696]: I0321 08:40:49.904397 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkqs9\" (UniqueName: \"kubernetes.io/projected/f46ecbe7-f1ca-4d24-975f-487b15fb2707-kube-api-access-mkqs9\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8\" (UID: \"f46ecbe7-f1ca-4d24-975f-487b15fb2707\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8" Mar 21 08:40:49 crc kubenswrapper[4696]: I0321 08:40:49.904677 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f46ecbe7-f1ca-4d24-975f-487b15fb2707-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8\" (UID: \"f46ecbe7-f1ca-4d24-975f-487b15fb2707\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8" Mar 21 08:40:49 crc kubenswrapper[4696]: I0321 08:40:49.904797 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f46ecbe7-f1ca-4d24-975f-487b15fb2707-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8\" (UID: \"f46ecbe7-f1ca-4d24-975f-487b15fb2707\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8" Mar 21 08:40:50 crc kubenswrapper[4696]: I0321 08:40:50.005890 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f46ecbe7-f1ca-4d24-975f-487b15fb2707-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8\" (UID: \"f46ecbe7-f1ca-4d24-975f-487b15fb2707\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8" Mar 21 08:40:50 crc kubenswrapper[4696]: I0321 08:40:50.005942 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkqs9\" (UniqueName: \"kubernetes.io/projected/f46ecbe7-f1ca-4d24-975f-487b15fb2707-kube-api-access-mkqs9\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8\" (UID: \"f46ecbe7-f1ca-4d24-975f-487b15fb2707\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8" Mar 21 08:40:50 crc kubenswrapper[4696]: I0321 08:40:50.005987 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f46ecbe7-f1ca-4d24-975f-487b15fb2707-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8\" (UID: \"f46ecbe7-f1ca-4d24-975f-487b15fb2707\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8" Mar 21 08:40:50 crc kubenswrapper[4696]: I0321 08:40:50.006389 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f46ecbe7-f1ca-4d24-975f-487b15fb2707-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8\" (UID: \"f46ecbe7-f1ca-4d24-975f-487b15fb2707\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8" Mar 21 08:40:50 crc kubenswrapper[4696]: I0321 08:40:50.007155 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f46ecbe7-f1ca-4d24-975f-487b15fb2707-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8\" (UID: \"f46ecbe7-f1ca-4d24-975f-487b15fb2707\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8" Mar 21 08:40:50 crc kubenswrapper[4696]: I0321 08:40:50.027078 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkqs9\" (UniqueName: \"kubernetes.io/projected/f46ecbe7-f1ca-4d24-975f-487b15fb2707-kube-api-access-mkqs9\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8\" (UID: \"f46ecbe7-f1ca-4d24-975f-487b15fb2707\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8" Mar 21 08:40:50 crc kubenswrapper[4696]: I0321 08:40:50.183862 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8" Mar 21 08:40:50 crc kubenswrapper[4696]: I0321 08:40:50.397416 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8"] Mar 21 08:40:51 crc kubenswrapper[4696]: I0321 08:40:51.236790 4696 generic.go:334] "Generic (PLEG): container finished" podID="f46ecbe7-f1ca-4d24-975f-487b15fb2707" containerID="4b9a02daaa80cb2798d9cdd1948b3e5310fb0da885ba1531f457087514d8bf09" exitCode=0 Mar 21 08:40:51 crc kubenswrapper[4696]: I0321 08:40:51.237014 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8" event={"ID":"f46ecbe7-f1ca-4d24-975f-487b15fb2707","Type":"ContainerDied","Data":"4b9a02daaa80cb2798d9cdd1948b3e5310fb0da885ba1531f457087514d8bf09"} Mar 21 08:40:51 crc kubenswrapper[4696]: I0321 08:40:51.237192 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8" event={"ID":"f46ecbe7-f1ca-4d24-975f-487b15fb2707","Type":"ContainerStarted","Data":"d436af1ac9bca94c7228fb9af0bca9a1e89550b9ef2b3872c39af521b4cce257"} Mar 21 08:40:53 crc kubenswrapper[4696]: I0321 08:40:53.251575 4696 generic.go:334] "Generic (PLEG): container finished" podID="f46ecbe7-f1ca-4d24-975f-487b15fb2707" containerID="880d14a7b2d1a68c35032fd026c4fa25a8ec74022977e0b69417787d44aad01e" exitCode=0 Mar 21 08:40:53 crc kubenswrapper[4696]: I0321 08:40:53.251634 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8" event={"ID":"f46ecbe7-f1ca-4d24-975f-487b15fb2707","Type":"ContainerDied","Data":"880d14a7b2d1a68c35032fd026c4fa25a8ec74022977e0b69417787d44aad01e"} Mar 21 08:40:54 crc kubenswrapper[4696]: I0321 08:40:54.260685 4696 generic.go:334] "Generic (PLEG): container finished" podID="f46ecbe7-f1ca-4d24-975f-487b15fb2707" containerID="40b01669b5726ec0c41314d765bca1e3de16b665a5fa7a1515926a1ba2be2c34" exitCode=0 Mar 21 08:40:54 crc kubenswrapper[4696]: I0321 08:40:54.260768 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8" event={"ID":"f46ecbe7-f1ca-4d24-975f-487b15fb2707","Type":"ContainerDied","Data":"40b01669b5726ec0c41314d765bca1e3de16b665a5fa7a1515926a1ba2be2c34"} Mar 21 08:40:55 crc kubenswrapper[4696]: I0321 08:40:55.533418 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8" Mar 21 08:40:55 crc kubenswrapper[4696]: I0321 08:40:55.580301 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkqs9\" (UniqueName: \"kubernetes.io/projected/f46ecbe7-f1ca-4d24-975f-487b15fb2707-kube-api-access-mkqs9\") pod \"f46ecbe7-f1ca-4d24-975f-487b15fb2707\" (UID: \"f46ecbe7-f1ca-4d24-975f-487b15fb2707\") " Mar 21 08:40:55 crc kubenswrapper[4696]: I0321 08:40:55.580612 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f46ecbe7-f1ca-4d24-975f-487b15fb2707-bundle\") pod \"f46ecbe7-f1ca-4d24-975f-487b15fb2707\" (UID: \"f46ecbe7-f1ca-4d24-975f-487b15fb2707\") " Mar 21 08:40:55 crc kubenswrapper[4696]: I0321 08:40:55.580750 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f46ecbe7-f1ca-4d24-975f-487b15fb2707-util\") pod \"f46ecbe7-f1ca-4d24-975f-487b15fb2707\" (UID: \"f46ecbe7-f1ca-4d24-975f-487b15fb2707\") " Mar 21 08:40:55 crc kubenswrapper[4696]: I0321 08:40:55.581225 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f46ecbe7-f1ca-4d24-975f-487b15fb2707-bundle" (OuterVolumeSpecName: "bundle") pod "f46ecbe7-f1ca-4d24-975f-487b15fb2707" (UID: "f46ecbe7-f1ca-4d24-975f-487b15fb2707"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:40:55 crc kubenswrapper[4696]: I0321 08:40:55.595453 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f46ecbe7-f1ca-4d24-975f-487b15fb2707-util" (OuterVolumeSpecName: "util") pod "f46ecbe7-f1ca-4d24-975f-487b15fb2707" (UID: "f46ecbe7-f1ca-4d24-975f-487b15fb2707"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:40:55 crc kubenswrapper[4696]: I0321 08:40:55.602618 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f46ecbe7-f1ca-4d24-975f-487b15fb2707-kube-api-access-mkqs9" (OuterVolumeSpecName: "kube-api-access-mkqs9") pod "f46ecbe7-f1ca-4d24-975f-487b15fb2707" (UID: "f46ecbe7-f1ca-4d24-975f-487b15fb2707"). InnerVolumeSpecName "kube-api-access-mkqs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:40:55 crc kubenswrapper[4696]: I0321 08:40:55.681944 4696 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f46ecbe7-f1ca-4d24-975f-487b15fb2707-util\") on node \"crc\" DevicePath \"\"" Mar 21 08:40:55 crc kubenswrapper[4696]: I0321 08:40:55.681974 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkqs9\" (UniqueName: \"kubernetes.io/projected/f46ecbe7-f1ca-4d24-975f-487b15fb2707-kube-api-access-mkqs9\") on node \"crc\" DevicePath \"\"" Mar 21 08:40:55 crc kubenswrapper[4696]: I0321 08:40:55.681986 4696 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f46ecbe7-f1ca-4d24-975f-487b15fb2707-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:40:56 crc kubenswrapper[4696]: I0321 08:40:56.273159 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8" event={"ID":"f46ecbe7-f1ca-4d24-975f-487b15fb2707","Type":"ContainerDied","Data":"d436af1ac9bca94c7228fb9af0bca9a1e89550b9ef2b3872c39af521b4cce257"} Mar 21 08:40:56 crc kubenswrapper[4696]: I0321 08:40:56.273210 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d436af1ac9bca94c7228fb9af0bca9a1e89550b9ef2b3872c39af521b4cce257" Mar 21 08:40:56 crc kubenswrapper[4696]: I0321 08:40:56.273272 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8" Mar 21 08:40:59 crc kubenswrapper[4696]: I0321 08:40:59.337832 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-zc76d"] Mar 21 08:40:59 crc kubenswrapper[4696]: E0321 08:40:59.338396 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f46ecbe7-f1ca-4d24-975f-487b15fb2707" containerName="extract" Mar 21 08:40:59 crc kubenswrapper[4696]: I0321 08:40:59.338413 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f46ecbe7-f1ca-4d24-975f-487b15fb2707" containerName="extract" Mar 21 08:40:59 crc kubenswrapper[4696]: E0321 08:40:59.338435 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f46ecbe7-f1ca-4d24-975f-487b15fb2707" containerName="util" Mar 21 08:40:59 crc kubenswrapper[4696]: I0321 08:40:59.338444 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f46ecbe7-f1ca-4d24-975f-487b15fb2707" containerName="util" Mar 21 08:40:59 crc kubenswrapper[4696]: E0321 08:40:59.338456 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f46ecbe7-f1ca-4d24-975f-487b15fb2707" containerName="pull" Mar 21 08:40:59 crc kubenswrapper[4696]: I0321 08:40:59.338464 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f46ecbe7-f1ca-4d24-975f-487b15fb2707" containerName="pull" Mar 21 08:40:59 crc kubenswrapper[4696]: I0321 08:40:59.338594 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f46ecbe7-f1ca-4d24-975f-487b15fb2707" containerName="extract" Mar 21 08:40:59 crc kubenswrapper[4696]: I0321 08:40:59.339083 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-zc76d" Mar 21 08:40:59 crc kubenswrapper[4696]: I0321 08:40:59.341864 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-zkdnl" Mar 21 08:40:59 crc kubenswrapper[4696]: I0321 08:40:59.341879 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Mar 21 08:40:59 crc kubenswrapper[4696]: I0321 08:40:59.342246 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Mar 21 08:40:59 crc kubenswrapper[4696]: I0321 08:40:59.354240 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-zc76d"] Mar 21 08:40:59 crc kubenswrapper[4696]: I0321 08:40:59.426455 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkl64\" (UniqueName: \"kubernetes.io/projected/3e70f161-07ef-4ed3-90be-6930cfb666a4-kube-api-access-xkl64\") pod \"nmstate-operator-796d4cfff4-zc76d\" (UID: \"3e70f161-07ef-4ed3-90be-6930cfb666a4\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-zc76d" Mar 21 08:40:59 crc kubenswrapper[4696]: I0321 08:40:59.527797 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkl64\" (UniqueName: \"kubernetes.io/projected/3e70f161-07ef-4ed3-90be-6930cfb666a4-kube-api-access-xkl64\") pod \"nmstate-operator-796d4cfff4-zc76d\" (UID: \"3e70f161-07ef-4ed3-90be-6930cfb666a4\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-zc76d" Mar 21 08:40:59 crc kubenswrapper[4696]: I0321 08:40:59.550032 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkl64\" (UniqueName: \"kubernetes.io/projected/3e70f161-07ef-4ed3-90be-6930cfb666a4-kube-api-access-xkl64\") pod \"nmstate-operator-796d4cfff4-zc76d\" (UID: \"3e70f161-07ef-4ed3-90be-6930cfb666a4\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-zc76d" Mar 21 08:40:59 crc kubenswrapper[4696]: I0321 08:40:59.657812 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-zc76d" Mar 21 08:41:00 crc kubenswrapper[4696]: I0321 08:41:00.065906 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-zc76d"] Mar 21 08:41:00 crc kubenswrapper[4696]: I0321 08:41:00.296227 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-zc76d" event={"ID":"3e70f161-07ef-4ed3-90be-6930cfb666a4","Type":"ContainerStarted","Data":"885b4257c8f4691418563ccaf348a09ab13889a4e54f97770f8a9dd3f693b9d1"} Mar 21 08:41:00 crc kubenswrapper[4696]: I0321 08:41:00.341162 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:41:00 crc kubenswrapper[4696]: I0321 08:41:00.341228 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:41:02 crc kubenswrapper[4696]: I0321 08:41:02.308143 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-zc76d" event={"ID":"3e70f161-07ef-4ed3-90be-6930cfb666a4","Type":"ContainerStarted","Data":"18727a687963bc397ec941d790c7c555e3b0d78a004bb87359dfd504d1113550"} Mar 21 08:41:02 crc kubenswrapper[4696]: I0321 08:41:02.322647 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-796d4cfff4-zc76d" podStartSLOduration=1.253460115 podStartE2EDuration="3.322626455s" podCreationTimestamp="2026-03-21 08:40:59 +0000 UTC" firstStartedPulling="2026-03-21 08:41:00.071101364 +0000 UTC m=+794.191982077" lastFinishedPulling="2026-03-21 08:41:02.140267704 +0000 UTC m=+796.261148417" observedRunningTime="2026-03-21 08:41:02.320472974 +0000 UTC m=+796.441353687" watchObservedRunningTime="2026-03-21 08:41:02.322626455 +0000 UTC m=+796.443507168" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.286884 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-q8tc5"] Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.288007 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-q8tc5" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.290099 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-g46sx" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.299689 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-q8tc5"] Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.305159 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-q8s95"] Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.305796 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-q8s95" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.309593 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.323767 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-r8wnr"] Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.324684 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-r8wnr" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.336909 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-q8s95"] Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.377364 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpxj2\" (UniqueName: \"kubernetes.io/projected/a351e45b-b9be-4898-8e07-c63cdd497245-kube-api-access-bpxj2\") pod \"nmstate-webhook-5f558f5558-q8s95\" (UID: \"a351e45b-b9be-4898-8e07-c63cdd497245\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-q8s95" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.377420 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/be6dcd14-1d77-4c67-b527-29e7f46efc3e-nmstate-lock\") pod \"nmstate-handler-r8wnr\" (UID: \"be6dcd14-1d77-4c67-b527-29e7f46efc3e\") " pod="openshift-nmstate/nmstate-handler-r8wnr" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.377445 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zghh9\" (UniqueName: \"kubernetes.io/projected/be6dcd14-1d77-4c67-b527-29e7f46efc3e-kube-api-access-zghh9\") pod \"nmstate-handler-r8wnr\" (UID: \"be6dcd14-1d77-4c67-b527-29e7f46efc3e\") " pod="openshift-nmstate/nmstate-handler-r8wnr" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.377473 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfhg2\" (UniqueName: \"kubernetes.io/projected/078ef7e2-dd73-486e-9213-804f5dbd26b2-kube-api-access-zfhg2\") pod \"nmstate-metrics-9b8c8685d-q8tc5\" (UID: \"078ef7e2-dd73-486e-9213-804f5dbd26b2\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-q8tc5" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.377497 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/a351e45b-b9be-4898-8e07-c63cdd497245-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-q8s95\" (UID: \"a351e45b-b9be-4898-8e07-c63cdd497245\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-q8s95" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.377550 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/be6dcd14-1d77-4c67-b527-29e7f46efc3e-ovs-socket\") pod \"nmstate-handler-r8wnr\" (UID: \"be6dcd14-1d77-4c67-b527-29e7f46efc3e\") " pod="openshift-nmstate/nmstate-handler-r8wnr" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.377572 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/be6dcd14-1d77-4c67-b527-29e7f46efc3e-dbus-socket\") pod \"nmstate-handler-r8wnr\" (UID: \"be6dcd14-1d77-4c67-b527-29e7f46efc3e\") " pod="openshift-nmstate/nmstate-handler-r8wnr" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.424786 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-k9rdf"] Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.425541 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-k9rdf" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.429707 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.429710 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.430811 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-4zwf6" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.447175 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-k9rdf"] Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.478900 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/a351e45b-b9be-4898-8e07-c63cdd497245-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-q8s95\" (UID: \"a351e45b-b9be-4898-8e07-c63cdd497245\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-q8s95" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.478972 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/be6dcd14-1d77-4c67-b527-29e7f46efc3e-ovs-socket\") pod \"nmstate-handler-r8wnr\" (UID: \"be6dcd14-1d77-4c67-b527-29e7f46efc3e\") " pod="openshift-nmstate/nmstate-handler-r8wnr" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.479002 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/be6dcd14-1d77-4c67-b527-29e7f46efc3e-dbus-socket\") pod \"nmstate-handler-r8wnr\" (UID: \"be6dcd14-1d77-4c67-b527-29e7f46efc3e\") " pod="openshift-nmstate/nmstate-handler-r8wnr" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.479026 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-k9rdf\" (UID: \"81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-k9rdf" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.479064 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpxj2\" (UniqueName: \"kubernetes.io/projected/a351e45b-b9be-4898-8e07-c63cdd497245-kube-api-access-bpxj2\") pod \"nmstate-webhook-5f558f5558-q8s95\" (UID: \"a351e45b-b9be-4898-8e07-c63cdd497245\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-q8s95" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.479085 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-k9rdf\" (UID: \"81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-k9rdf" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.479118 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/be6dcd14-1d77-4c67-b527-29e7f46efc3e-nmstate-lock\") pod \"nmstate-handler-r8wnr\" (UID: \"be6dcd14-1d77-4c67-b527-29e7f46efc3e\") " pod="openshift-nmstate/nmstate-handler-r8wnr" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.479142 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zghh9\" (UniqueName: \"kubernetes.io/projected/be6dcd14-1d77-4c67-b527-29e7f46efc3e-kube-api-access-zghh9\") pod \"nmstate-handler-r8wnr\" (UID: \"be6dcd14-1d77-4c67-b527-29e7f46efc3e\") " pod="openshift-nmstate/nmstate-handler-r8wnr" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.479165 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfhg2\" (UniqueName: \"kubernetes.io/projected/078ef7e2-dd73-486e-9213-804f5dbd26b2-kube-api-access-zfhg2\") pod \"nmstate-metrics-9b8c8685d-q8tc5\" (UID: \"078ef7e2-dd73-486e-9213-804f5dbd26b2\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-q8tc5" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.479189 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-922fl\" (UniqueName: \"kubernetes.io/projected/81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160-kube-api-access-922fl\") pod \"nmstate-console-plugin-86f58fcf4-k9rdf\" (UID: \"81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-k9rdf" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.479261 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/be6dcd14-1d77-4c67-b527-29e7f46efc3e-ovs-socket\") pod \"nmstate-handler-r8wnr\" (UID: \"be6dcd14-1d77-4c67-b527-29e7f46efc3e\") " pod="openshift-nmstate/nmstate-handler-r8wnr" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.479377 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/be6dcd14-1d77-4c67-b527-29e7f46efc3e-nmstate-lock\") pod \"nmstate-handler-r8wnr\" (UID: \"be6dcd14-1d77-4c67-b527-29e7f46efc3e\") " pod="openshift-nmstate/nmstate-handler-r8wnr" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.479609 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/be6dcd14-1d77-4c67-b527-29e7f46efc3e-dbus-socket\") pod \"nmstate-handler-r8wnr\" (UID: \"be6dcd14-1d77-4c67-b527-29e7f46efc3e\") " pod="openshift-nmstate/nmstate-handler-r8wnr" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.494724 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/a351e45b-b9be-4898-8e07-c63cdd497245-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-q8s95\" (UID: \"a351e45b-b9be-4898-8e07-c63cdd497245\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-q8s95" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.497585 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfhg2\" (UniqueName: \"kubernetes.io/projected/078ef7e2-dd73-486e-9213-804f5dbd26b2-kube-api-access-zfhg2\") pod \"nmstate-metrics-9b8c8685d-q8tc5\" (UID: \"078ef7e2-dd73-486e-9213-804f5dbd26b2\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-q8tc5" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.502416 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpxj2\" (UniqueName: \"kubernetes.io/projected/a351e45b-b9be-4898-8e07-c63cdd497245-kube-api-access-bpxj2\") pod \"nmstate-webhook-5f558f5558-q8s95\" (UID: \"a351e45b-b9be-4898-8e07-c63cdd497245\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-q8s95" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.504985 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zghh9\" (UniqueName: \"kubernetes.io/projected/be6dcd14-1d77-4c67-b527-29e7f46efc3e-kube-api-access-zghh9\") pod \"nmstate-handler-r8wnr\" (UID: \"be6dcd14-1d77-4c67-b527-29e7f46efc3e\") " pod="openshift-nmstate/nmstate-handler-r8wnr" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.580486 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-k9rdf\" (UID: \"81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-k9rdf" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.580572 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-922fl\" (UniqueName: \"kubernetes.io/projected/81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160-kube-api-access-922fl\") pod \"nmstate-console-plugin-86f58fcf4-k9rdf\" (UID: \"81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-k9rdf" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.580646 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-k9rdf\" (UID: \"81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-k9rdf" Mar 21 08:41:03 crc kubenswrapper[4696]: E0321 08:41:03.580769 4696 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Mar 21 08:41:03 crc kubenswrapper[4696]: E0321 08:41:03.580856 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160-plugin-serving-cert podName:81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160 nodeName:}" failed. No retries permitted until 2026-03-21 08:41:04.080832753 +0000 UTC m=+798.201713466 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160-plugin-serving-cert") pod "nmstate-console-plugin-86f58fcf4-k9rdf" (UID: "81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160") : secret "plugin-serving-cert" not found Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.581351 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-k9rdf\" (UID: \"81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-k9rdf" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.598699 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-922fl\" (UniqueName: \"kubernetes.io/projected/81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160-kube-api-access-922fl\") pod \"nmstate-console-plugin-86f58fcf4-k9rdf\" (UID: \"81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-k9rdf" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.613674 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-q8tc5" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.626148 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-578df8f64b-87v7q"] Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.626677 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-q8s95" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.627163 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.631086 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-578df8f64b-87v7q"] Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.671671 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-r8wnr" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.681467 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4e08dc8f-cc6b-400e-b49a-85f12df04d45-console-serving-cert\") pod \"console-578df8f64b-87v7q\" (UID: \"4e08dc8f-cc6b-400e-b49a-85f12df04d45\") " pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.681515 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4e08dc8f-cc6b-400e-b49a-85f12df04d45-console-oauth-config\") pod \"console-578df8f64b-87v7q\" (UID: \"4e08dc8f-cc6b-400e-b49a-85f12df04d45\") " pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.681566 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4e08dc8f-cc6b-400e-b49a-85f12df04d45-console-config\") pod \"console-578df8f64b-87v7q\" (UID: \"4e08dc8f-cc6b-400e-b49a-85f12df04d45\") " pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.681593 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4e08dc8f-cc6b-400e-b49a-85f12df04d45-service-ca\") pod \"console-578df8f64b-87v7q\" (UID: \"4e08dc8f-cc6b-400e-b49a-85f12df04d45\") " pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.681630 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4e08dc8f-cc6b-400e-b49a-85f12df04d45-oauth-serving-cert\") pod \"console-578df8f64b-87v7q\" (UID: \"4e08dc8f-cc6b-400e-b49a-85f12df04d45\") " pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.681675 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4e08dc8f-cc6b-400e-b49a-85f12df04d45-trusted-ca-bundle\") pod \"console-578df8f64b-87v7q\" (UID: \"4e08dc8f-cc6b-400e-b49a-85f12df04d45\") " pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.681692 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgfvq\" (UniqueName: \"kubernetes.io/projected/4e08dc8f-cc6b-400e-b49a-85f12df04d45-kube-api-access-vgfvq\") pod \"console-578df8f64b-87v7q\" (UID: \"4e08dc8f-cc6b-400e-b49a-85f12df04d45\") " pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: W0321 08:41:03.692900 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe6dcd14_1d77_4c67_b527_29e7f46efc3e.slice/crio-f8ccb3fc3b5c333e1f78bee466ba6ac810ab069ff2d0926dcc8eddce00d31c09 WatchSource:0}: Error finding container f8ccb3fc3b5c333e1f78bee466ba6ac810ab069ff2d0926dcc8eddce00d31c09: Status 404 returned error can't find the container with id f8ccb3fc3b5c333e1f78bee466ba6ac810ab069ff2d0926dcc8eddce00d31c09 Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.698483 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.784370 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4e08dc8f-cc6b-400e-b49a-85f12df04d45-console-oauth-config\") pod \"console-578df8f64b-87v7q\" (UID: \"4e08dc8f-cc6b-400e-b49a-85f12df04d45\") " pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.784715 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4e08dc8f-cc6b-400e-b49a-85f12df04d45-console-config\") pod \"console-578df8f64b-87v7q\" (UID: \"4e08dc8f-cc6b-400e-b49a-85f12df04d45\") " pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.784730 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4e08dc8f-cc6b-400e-b49a-85f12df04d45-service-ca\") pod \"console-578df8f64b-87v7q\" (UID: \"4e08dc8f-cc6b-400e-b49a-85f12df04d45\") " pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.784757 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4e08dc8f-cc6b-400e-b49a-85f12df04d45-oauth-serving-cert\") pod \"console-578df8f64b-87v7q\" (UID: \"4e08dc8f-cc6b-400e-b49a-85f12df04d45\") " pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.784781 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4e08dc8f-cc6b-400e-b49a-85f12df04d45-trusted-ca-bundle\") pod \"console-578df8f64b-87v7q\" (UID: \"4e08dc8f-cc6b-400e-b49a-85f12df04d45\") " pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.784796 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgfvq\" (UniqueName: \"kubernetes.io/projected/4e08dc8f-cc6b-400e-b49a-85f12df04d45-kube-api-access-vgfvq\") pod \"console-578df8f64b-87v7q\" (UID: \"4e08dc8f-cc6b-400e-b49a-85f12df04d45\") " pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.784878 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4e08dc8f-cc6b-400e-b49a-85f12df04d45-console-serving-cert\") pod \"console-578df8f64b-87v7q\" (UID: \"4e08dc8f-cc6b-400e-b49a-85f12df04d45\") " pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.787719 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4e08dc8f-cc6b-400e-b49a-85f12df04d45-service-ca\") pod \"console-578df8f64b-87v7q\" (UID: \"4e08dc8f-cc6b-400e-b49a-85f12df04d45\") " pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.788671 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4e08dc8f-cc6b-400e-b49a-85f12df04d45-oauth-serving-cert\") pod \"console-578df8f64b-87v7q\" (UID: \"4e08dc8f-cc6b-400e-b49a-85f12df04d45\") " pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.788684 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4e08dc8f-cc6b-400e-b49a-85f12df04d45-console-config\") pod \"console-578df8f64b-87v7q\" (UID: \"4e08dc8f-cc6b-400e-b49a-85f12df04d45\") " pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.789240 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4e08dc8f-cc6b-400e-b49a-85f12df04d45-trusted-ca-bundle\") pod \"console-578df8f64b-87v7q\" (UID: \"4e08dc8f-cc6b-400e-b49a-85f12df04d45\") " pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.790265 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4e08dc8f-cc6b-400e-b49a-85f12df04d45-console-serving-cert\") pod \"console-578df8f64b-87v7q\" (UID: \"4e08dc8f-cc6b-400e-b49a-85f12df04d45\") " pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.792367 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4e08dc8f-cc6b-400e-b49a-85f12df04d45-console-oauth-config\") pod \"console-578df8f64b-87v7q\" (UID: \"4e08dc8f-cc6b-400e-b49a-85f12df04d45\") " pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.802484 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgfvq\" (UniqueName: \"kubernetes.io/projected/4e08dc8f-cc6b-400e-b49a-85f12df04d45-kube-api-access-vgfvq\") pod \"console-578df8f64b-87v7q\" (UID: \"4e08dc8f-cc6b-400e-b49a-85f12df04d45\") " pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.836979 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-q8tc5"] Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.865215 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-q8s95"] Mar 21 08:41:03 crc kubenswrapper[4696]: W0321 08:41:03.868707 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda351e45b_b9be_4898_8e07_c63cdd497245.slice/crio-e4c13719db0968c49995dca51899b97494f44216f3f9ad22eee5417bdad8c947 WatchSource:0}: Error finding container e4c13719db0968c49995dca51899b97494f44216f3f9ad22eee5417bdad8c947: Status 404 returned error can't find the container with id e4c13719db0968c49995dca51899b97494f44216f3f9ad22eee5417bdad8c947 Mar 21 08:41:03 crc kubenswrapper[4696]: I0321 08:41:03.987133 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:04 crc kubenswrapper[4696]: I0321 08:41:04.088055 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-k9rdf\" (UID: \"81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-k9rdf" Mar 21 08:41:04 crc kubenswrapper[4696]: I0321 08:41:04.090707 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-k9rdf\" (UID: \"81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-k9rdf" Mar 21 08:41:04 crc kubenswrapper[4696]: I0321 08:41:04.169785 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-578df8f64b-87v7q"] Mar 21 08:41:04 crc kubenswrapper[4696]: I0321 08:41:04.321885 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-q8s95" event={"ID":"a351e45b-b9be-4898-8e07-c63cdd497245","Type":"ContainerStarted","Data":"e4c13719db0968c49995dca51899b97494f44216f3f9ad22eee5417bdad8c947"} Mar 21 08:41:04 crc kubenswrapper[4696]: I0321 08:41:04.323229 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-r8wnr" event={"ID":"be6dcd14-1d77-4c67-b527-29e7f46efc3e","Type":"ContainerStarted","Data":"f8ccb3fc3b5c333e1f78bee466ba6ac810ab069ff2d0926dcc8eddce00d31c09"} Mar 21 08:41:04 crc kubenswrapper[4696]: I0321 08:41:04.324322 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-q8tc5" event={"ID":"078ef7e2-dd73-486e-9213-804f5dbd26b2","Type":"ContainerStarted","Data":"0d7583f853a1d17df91c591dfca120e21814e483d3be03325fe71054e79c2a29"} Mar 21 08:41:04 crc kubenswrapper[4696]: I0321 08:41:04.325873 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-578df8f64b-87v7q" event={"ID":"4e08dc8f-cc6b-400e-b49a-85f12df04d45","Type":"ContainerStarted","Data":"6b07e4f3cfbcc202cf5e548c446a046ed209acf2b34493ee19301984f52676aa"} Mar 21 08:41:04 crc kubenswrapper[4696]: I0321 08:41:04.325997 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-578df8f64b-87v7q" event={"ID":"4e08dc8f-cc6b-400e-b49a-85f12df04d45","Type":"ContainerStarted","Data":"2b9aa2b8d9409295a0c1da6b2d57480d9fd19f78b565f473a744165b05f504d0"} Mar 21 08:41:04 crc kubenswrapper[4696]: I0321 08:41:04.337417 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-k9rdf" Mar 21 08:41:04 crc kubenswrapper[4696]: I0321 08:41:04.351945 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-578df8f64b-87v7q" podStartSLOduration=1.351928797 podStartE2EDuration="1.351928797s" podCreationTimestamp="2026-03-21 08:41:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:41:04.351337094 +0000 UTC m=+798.472217817" watchObservedRunningTime="2026-03-21 08:41:04.351928797 +0000 UTC m=+798.472809510" Mar 21 08:41:04 crc kubenswrapper[4696]: I0321 08:41:04.788161 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-k9rdf"] Mar 21 08:41:04 crc kubenswrapper[4696]: W0321 08:41:04.790664 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81abfcbd_f8b8_48d5_99b2_3cb7cc2d9160.slice/crio-01daedc8424f0ab0209a86c811d2bfdd28c2e36aae868f7803f12380b71c49e2 WatchSource:0}: Error finding container 01daedc8424f0ab0209a86c811d2bfdd28c2e36aae868f7803f12380b71c49e2: Status 404 returned error can't find the container with id 01daedc8424f0ab0209a86c811d2bfdd28c2e36aae868f7803f12380b71c49e2 Mar 21 08:41:05 crc kubenswrapper[4696]: I0321 08:41:05.355925 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-k9rdf" event={"ID":"81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160","Type":"ContainerStarted","Data":"01daedc8424f0ab0209a86c811d2bfdd28c2e36aae868f7803f12380b71c49e2"} Mar 21 08:41:07 crc kubenswrapper[4696]: I0321 08:41:07.368436 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-q8tc5" event={"ID":"078ef7e2-dd73-486e-9213-804f5dbd26b2","Type":"ContainerStarted","Data":"3c94f70b6a324f59c83918398907088fc15766533fe99e28461fd9d684578ad6"} Mar 21 08:41:07 crc kubenswrapper[4696]: I0321 08:41:07.370770 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-q8s95" event={"ID":"a351e45b-b9be-4898-8e07-c63cdd497245","Type":"ContainerStarted","Data":"c15a7fc8a8695dc79bf0080ce130eccd1f36e02b695836c94bc221c173813fcc"} Mar 21 08:41:07 crc kubenswrapper[4696]: I0321 08:41:07.371857 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f558f5558-q8s95" Mar 21 08:41:07 crc kubenswrapper[4696]: I0321 08:41:07.373491 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-r8wnr" event={"ID":"be6dcd14-1d77-4c67-b527-29e7f46efc3e","Type":"ContainerStarted","Data":"966fd5b50e11821918a3b8854ef4f123150d89fab566084379d792f59f94977e"} Mar 21 08:41:07 crc kubenswrapper[4696]: I0321 08:41:07.374091 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-r8wnr" Mar 21 08:41:07 crc kubenswrapper[4696]: I0321 08:41:07.393830 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f558f5558-q8s95" podStartSLOduration=1.807196332 podStartE2EDuration="4.393796298s" podCreationTimestamp="2026-03-21 08:41:03 +0000 UTC" firstStartedPulling="2026-03-21 08:41:03.870860128 +0000 UTC m=+797.991740841" lastFinishedPulling="2026-03-21 08:41:06.457460094 +0000 UTC m=+800.578340807" observedRunningTime="2026-03-21 08:41:07.387719305 +0000 UTC m=+801.508600038" watchObservedRunningTime="2026-03-21 08:41:07.393796298 +0000 UTC m=+801.514677011" Mar 21 08:41:08 crc kubenswrapper[4696]: I0321 08:41:08.384583 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-k9rdf" event={"ID":"81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160","Type":"ContainerStarted","Data":"eabf783fe400a651cf5a9ed08331f98bd2651aed04d1bb0add4ab503dd3bf24f"} Mar 21 08:41:08 crc kubenswrapper[4696]: I0321 08:41:08.398740 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-k9rdf" podStartSLOduration=2.773171272 podStartE2EDuration="5.398727306s" podCreationTimestamp="2026-03-21 08:41:03 +0000 UTC" firstStartedPulling="2026-03-21 08:41:04.79400267 +0000 UTC m=+798.914883383" lastFinishedPulling="2026-03-21 08:41:07.419558694 +0000 UTC m=+801.540439417" observedRunningTime="2026-03-21 08:41:08.397588729 +0000 UTC m=+802.518469442" watchObservedRunningTime="2026-03-21 08:41:08.398727306 +0000 UTC m=+802.519608019" Mar 21 08:41:08 crc kubenswrapper[4696]: I0321 08:41:08.401105 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-r8wnr" podStartSLOduration=2.6783438889999998 podStartE2EDuration="5.40107918s" podCreationTimestamp="2026-03-21 08:41:03 +0000 UTC" firstStartedPulling="2026-03-21 08:41:03.698242405 +0000 UTC m=+797.819123118" lastFinishedPulling="2026-03-21 08:41:06.420977696 +0000 UTC m=+800.541858409" observedRunningTime="2026-03-21 08:41:07.412761514 +0000 UTC m=+801.533642237" watchObservedRunningTime="2026-03-21 08:41:08.40107918 +0000 UTC m=+802.521959913" Mar 21 08:41:09 crc kubenswrapper[4696]: I0321 08:41:09.393615 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-q8tc5" event={"ID":"078ef7e2-dd73-486e-9213-804f5dbd26b2","Type":"ContainerStarted","Data":"42b45a85470664917ba90467dd2f4c253ea5721c7b3acb9ee4a79aa02c6eefd7"} Mar 21 08:41:09 crc kubenswrapper[4696]: I0321 08:41:09.406805 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-q8tc5" podStartSLOduration=1.3564453539999999 podStartE2EDuration="6.406779307s" podCreationTimestamp="2026-03-21 08:41:03 +0000 UTC" firstStartedPulling="2026-03-21 08:41:03.849097595 +0000 UTC m=+797.969978308" lastFinishedPulling="2026-03-21 08:41:08.899431548 +0000 UTC m=+803.020312261" observedRunningTime="2026-03-21 08:41:09.406254124 +0000 UTC m=+803.527134827" watchObservedRunningTime="2026-03-21 08:41:09.406779307 +0000 UTC m=+803.527660040" Mar 21 08:41:13 crc kubenswrapper[4696]: I0321 08:41:13.696561 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-r8wnr" Mar 21 08:41:13 crc kubenswrapper[4696]: I0321 08:41:13.987551 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:13 crc kubenswrapper[4696]: I0321 08:41:13.987874 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:13 crc kubenswrapper[4696]: I0321 08:41:13.991463 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:14 crc kubenswrapper[4696]: I0321 08:41:14.426125 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-578df8f64b-87v7q" Mar 21 08:41:14 crc kubenswrapper[4696]: I0321 08:41:14.491482 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-f5n2g"] Mar 21 08:41:23 crc kubenswrapper[4696]: I0321 08:41:23.633568 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f558f5558-q8s95" Mar 21 08:41:30 crc kubenswrapper[4696]: I0321 08:41:30.341800 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:41:30 crc kubenswrapper[4696]: I0321 08:41:30.342395 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:41:30 crc kubenswrapper[4696]: I0321 08:41:30.342458 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:41:30 crc kubenswrapper[4696]: I0321 08:41:30.343133 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a451c8e7ce2e953d0eead4e37b6cb08fe2df64dae3d013237676d9b8ff9dbb40"} pod="openshift-machine-config-operator/machine-config-daemon-z7srw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 08:41:30 crc kubenswrapper[4696]: I0321 08:41:30.343203 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" containerID="cri-o://a451c8e7ce2e953d0eead4e37b6cb08fe2df64dae3d013237676d9b8ff9dbb40" gracePeriod=600 Mar 21 08:41:30 crc kubenswrapper[4696]: I0321 08:41:30.543934 4696 generic.go:334] "Generic (PLEG): container finished" podID="daaf227a-2305-495e-8495-a280abcd8e10" containerID="a451c8e7ce2e953d0eead4e37b6cb08fe2df64dae3d013237676d9b8ff9dbb40" exitCode=0 Mar 21 08:41:30 crc kubenswrapper[4696]: I0321 08:41:30.545582 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerDied","Data":"a451c8e7ce2e953d0eead4e37b6cb08fe2df64dae3d013237676d9b8ff9dbb40"} Mar 21 08:41:30 crc kubenswrapper[4696]: I0321 08:41:30.545671 4696 scope.go:117] "RemoveContainer" containerID="fc18c7d093fe04a689c494e8d2c8a4d315eb2326b2edffdf3f61d7449dee30d2" Mar 21 08:41:31 crc kubenswrapper[4696]: I0321 08:41:31.554364 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerStarted","Data":"06727ab615d47d51e86a72ca06fb67dd4f0d596dbb88f8ea56eca5339ae5cc33"} Mar 21 08:41:38 crc kubenswrapper[4696]: I0321 08:41:38.573115 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9"] Mar 21 08:41:38 crc kubenswrapper[4696]: I0321 08:41:38.574649 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9" Mar 21 08:41:38 crc kubenswrapper[4696]: I0321 08:41:38.576520 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 21 08:41:38 crc kubenswrapper[4696]: I0321 08:41:38.588159 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9"] Mar 21 08:41:38 crc kubenswrapper[4696]: I0321 08:41:38.665784 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnkv2\" (UniqueName: \"kubernetes.io/projected/6ed1f144-c0f4-470e-9bc2-fc4ea36bc158-kube-api-access-lnkv2\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9\" (UID: \"6ed1f144-c0f4-470e-9bc2-fc4ea36bc158\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9" Mar 21 08:41:38 crc kubenswrapper[4696]: I0321 08:41:38.665854 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6ed1f144-c0f4-470e-9bc2-fc4ea36bc158-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9\" (UID: \"6ed1f144-c0f4-470e-9bc2-fc4ea36bc158\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9" Mar 21 08:41:38 crc kubenswrapper[4696]: I0321 08:41:38.665919 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6ed1f144-c0f4-470e-9bc2-fc4ea36bc158-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9\" (UID: \"6ed1f144-c0f4-470e-9bc2-fc4ea36bc158\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9" Mar 21 08:41:38 crc kubenswrapper[4696]: I0321 08:41:38.766907 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnkv2\" (UniqueName: \"kubernetes.io/projected/6ed1f144-c0f4-470e-9bc2-fc4ea36bc158-kube-api-access-lnkv2\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9\" (UID: \"6ed1f144-c0f4-470e-9bc2-fc4ea36bc158\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9" Mar 21 08:41:38 crc kubenswrapper[4696]: I0321 08:41:38.766972 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6ed1f144-c0f4-470e-9bc2-fc4ea36bc158-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9\" (UID: \"6ed1f144-c0f4-470e-9bc2-fc4ea36bc158\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9" Mar 21 08:41:38 crc kubenswrapper[4696]: I0321 08:41:38.767028 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6ed1f144-c0f4-470e-9bc2-fc4ea36bc158-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9\" (UID: \"6ed1f144-c0f4-470e-9bc2-fc4ea36bc158\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9" Mar 21 08:41:38 crc kubenswrapper[4696]: I0321 08:41:38.767521 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6ed1f144-c0f4-470e-9bc2-fc4ea36bc158-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9\" (UID: \"6ed1f144-c0f4-470e-9bc2-fc4ea36bc158\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9" Mar 21 08:41:38 crc kubenswrapper[4696]: I0321 08:41:38.767798 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6ed1f144-c0f4-470e-9bc2-fc4ea36bc158-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9\" (UID: \"6ed1f144-c0f4-470e-9bc2-fc4ea36bc158\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9" Mar 21 08:41:38 crc kubenswrapper[4696]: I0321 08:41:38.788260 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnkv2\" (UniqueName: \"kubernetes.io/projected/6ed1f144-c0f4-470e-9bc2-fc4ea36bc158-kube-api-access-lnkv2\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9\" (UID: \"6ed1f144-c0f4-470e-9bc2-fc4ea36bc158\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9" Mar 21 08:41:38 crc kubenswrapper[4696]: I0321 08:41:38.890583 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9" Mar 21 08:41:39 crc kubenswrapper[4696]: I0321 08:41:39.329598 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9"] Mar 21 08:41:39 crc kubenswrapper[4696]: I0321 08:41:39.540691 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-f5n2g" podUID="d1d05966-d105-4c45-bfa2-e66a67882baf" containerName="console" containerID="cri-o://35a8557dbdf27af515a9f6e30d7e692b787c70cfb3a58967964d4a12a9ccea12" gracePeriod=15 Mar 21 08:41:39 crc kubenswrapper[4696]: I0321 08:41:39.608012 4696 generic.go:334] "Generic (PLEG): container finished" podID="6ed1f144-c0f4-470e-9bc2-fc4ea36bc158" containerID="6f40772c34940b24612b4b7efd10074a48abc366c0324ef8b57954a51c590991" exitCode=0 Mar 21 08:41:39 crc kubenswrapper[4696]: I0321 08:41:39.608058 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9" event={"ID":"6ed1f144-c0f4-470e-9bc2-fc4ea36bc158","Type":"ContainerDied","Data":"6f40772c34940b24612b4b7efd10074a48abc366c0324ef8b57954a51c590991"} Mar 21 08:41:39 crc kubenswrapper[4696]: I0321 08:41:39.608087 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9" event={"ID":"6ed1f144-c0f4-470e-9bc2-fc4ea36bc158","Type":"ContainerStarted","Data":"de7359c4bb78fcd602b953a1f77681509968a6c23eccb438528c6b18eba0caa4"} Mar 21 08:41:39 crc kubenswrapper[4696]: I0321 08:41:39.869847 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-f5n2g_d1d05966-d105-4c45-bfa2-e66a67882baf/console/0.log" Mar 21 08:41:39 crc kubenswrapper[4696]: I0321 08:41:39.870185 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:41:39 crc kubenswrapper[4696]: I0321 08:41:39.982102 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-oauth-serving-cert\") pod \"d1d05966-d105-4c45-bfa2-e66a67882baf\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " Mar 21 08:41:39 crc kubenswrapper[4696]: I0321 08:41:39.982137 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-trusted-ca-bundle\") pod \"d1d05966-d105-4c45-bfa2-e66a67882baf\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " Mar 21 08:41:39 crc kubenswrapper[4696]: I0321 08:41:39.982235 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9q56\" (UniqueName: \"kubernetes.io/projected/d1d05966-d105-4c45-bfa2-e66a67882baf-kube-api-access-x9q56\") pod \"d1d05966-d105-4c45-bfa2-e66a67882baf\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " Mar 21 08:41:39 crc kubenswrapper[4696]: I0321 08:41:39.982290 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d1d05966-d105-4c45-bfa2-e66a67882baf-console-serving-cert\") pod \"d1d05966-d105-4c45-bfa2-e66a67882baf\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " Mar 21 08:41:39 crc kubenswrapper[4696]: I0321 08:41:39.982331 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-service-ca\") pod \"d1d05966-d105-4c45-bfa2-e66a67882baf\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " Mar 21 08:41:39 crc kubenswrapper[4696]: I0321 08:41:39.982418 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-console-config\") pod \"d1d05966-d105-4c45-bfa2-e66a67882baf\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " Mar 21 08:41:39 crc kubenswrapper[4696]: I0321 08:41:39.982482 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d1d05966-d105-4c45-bfa2-e66a67882baf-console-oauth-config\") pod \"d1d05966-d105-4c45-bfa2-e66a67882baf\" (UID: \"d1d05966-d105-4c45-bfa2-e66a67882baf\") " Mar 21 08:41:39 crc kubenswrapper[4696]: I0321 08:41:39.982960 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "d1d05966-d105-4c45-bfa2-e66a67882baf" (UID: "d1d05966-d105-4c45-bfa2-e66a67882baf"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:41:39 crc kubenswrapper[4696]: I0321 08:41:39.983485 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "d1d05966-d105-4c45-bfa2-e66a67882baf" (UID: "d1d05966-d105-4c45-bfa2-e66a67882baf"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:41:39 crc kubenswrapper[4696]: I0321 08:41:39.983493 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-console-config" (OuterVolumeSpecName: "console-config") pod "d1d05966-d105-4c45-bfa2-e66a67882baf" (UID: "d1d05966-d105-4c45-bfa2-e66a67882baf"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:41:39 crc kubenswrapper[4696]: I0321 08:41:39.983554 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-service-ca" (OuterVolumeSpecName: "service-ca") pod "d1d05966-d105-4c45-bfa2-e66a67882baf" (UID: "d1d05966-d105-4c45-bfa2-e66a67882baf"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:41:39 crc kubenswrapper[4696]: I0321 08:41:39.987606 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1d05966-d105-4c45-bfa2-e66a67882baf-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "d1d05966-d105-4c45-bfa2-e66a67882baf" (UID: "d1d05966-d105-4c45-bfa2-e66a67882baf"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:41:39 crc kubenswrapper[4696]: I0321 08:41:39.988156 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1d05966-d105-4c45-bfa2-e66a67882baf-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "d1d05966-d105-4c45-bfa2-e66a67882baf" (UID: "d1d05966-d105-4c45-bfa2-e66a67882baf"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:41:39 crc kubenswrapper[4696]: I0321 08:41:39.989097 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1d05966-d105-4c45-bfa2-e66a67882baf-kube-api-access-x9q56" (OuterVolumeSpecName: "kube-api-access-x9q56") pod "d1d05966-d105-4c45-bfa2-e66a67882baf" (UID: "d1d05966-d105-4c45-bfa2-e66a67882baf"). InnerVolumeSpecName "kube-api-access-x9q56". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:41:40 crc kubenswrapper[4696]: I0321 08:41:40.084001 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9q56\" (UniqueName: \"kubernetes.io/projected/d1d05966-d105-4c45-bfa2-e66a67882baf-kube-api-access-x9q56\") on node \"crc\" DevicePath \"\"" Mar 21 08:41:40 crc kubenswrapper[4696]: I0321 08:41:40.084041 4696 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d1d05966-d105-4c45-bfa2-e66a67882baf-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:41:40 crc kubenswrapper[4696]: I0321 08:41:40.084050 4696 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-service-ca\") on node \"crc\" DevicePath \"\"" Mar 21 08:41:40 crc kubenswrapper[4696]: I0321 08:41:40.084059 4696 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-console-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:41:40 crc kubenswrapper[4696]: I0321 08:41:40.084067 4696 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d1d05966-d105-4c45-bfa2-e66a67882baf-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:41:40 crc kubenswrapper[4696]: I0321 08:41:40.084074 4696 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 08:41:40 crc kubenswrapper[4696]: I0321 08:41:40.084082 4696 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d1d05966-d105-4c45-bfa2-e66a67882baf-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:41:40 crc kubenswrapper[4696]: I0321 08:41:40.617544 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-f5n2g_d1d05966-d105-4c45-bfa2-e66a67882baf/console/0.log" Mar 21 08:41:40 crc kubenswrapper[4696]: I0321 08:41:40.617595 4696 generic.go:334] "Generic (PLEG): container finished" podID="d1d05966-d105-4c45-bfa2-e66a67882baf" containerID="35a8557dbdf27af515a9f6e30d7e692b787c70cfb3a58967964d4a12a9ccea12" exitCode=2 Mar 21 08:41:40 crc kubenswrapper[4696]: I0321 08:41:40.617622 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-f5n2g" event={"ID":"d1d05966-d105-4c45-bfa2-e66a67882baf","Type":"ContainerDied","Data":"35a8557dbdf27af515a9f6e30d7e692b787c70cfb3a58967964d4a12a9ccea12"} Mar 21 08:41:40 crc kubenswrapper[4696]: I0321 08:41:40.617653 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-f5n2g" event={"ID":"d1d05966-d105-4c45-bfa2-e66a67882baf","Type":"ContainerDied","Data":"e8ba5226619f2620d146678c29fc7042c70e2dbf434d91f6483bbbba2262dc44"} Mar 21 08:41:40 crc kubenswrapper[4696]: I0321 08:41:40.617673 4696 scope.go:117] "RemoveContainer" containerID="35a8557dbdf27af515a9f6e30d7e692b787c70cfb3a58967964d4a12a9ccea12" Mar 21 08:41:40 crc kubenswrapper[4696]: I0321 08:41:40.617673 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-f5n2g" Mar 21 08:41:40 crc kubenswrapper[4696]: I0321 08:41:40.639012 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-f5n2g"] Mar 21 08:41:40 crc kubenswrapper[4696]: I0321 08:41:40.644528 4696 scope.go:117] "RemoveContainer" containerID="35a8557dbdf27af515a9f6e30d7e692b787c70cfb3a58967964d4a12a9ccea12" Mar 21 08:41:40 crc kubenswrapper[4696]: I0321 08:41:40.644715 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-f5n2g"] Mar 21 08:41:40 crc kubenswrapper[4696]: E0321 08:41:40.645159 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35a8557dbdf27af515a9f6e30d7e692b787c70cfb3a58967964d4a12a9ccea12\": container with ID starting with 35a8557dbdf27af515a9f6e30d7e692b787c70cfb3a58967964d4a12a9ccea12 not found: ID does not exist" containerID="35a8557dbdf27af515a9f6e30d7e692b787c70cfb3a58967964d4a12a9ccea12" Mar 21 08:41:40 crc kubenswrapper[4696]: I0321 08:41:40.645199 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35a8557dbdf27af515a9f6e30d7e692b787c70cfb3a58967964d4a12a9ccea12"} err="failed to get container status \"35a8557dbdf27af515a9f6e30d7e692b787c70cfb3a58967964d4a12a9ccea12\": rpc error: code = NotFound desc = could not find container \"35a8557dbdf27af515a9f6e30d7e692b787c70cfb3a58967964d4a12a9ccea12\": container with ID starting with 35a8557dbdf27af515a9f6e30d7e692b787c70cfb3a58967964d4a12a9ccea12 not found: ID does not exist" Mar 21 08:41:41 crc kubenswrapper[4696]: I0321 08:41:41.625622 4696 generic.go:334] "Generic (PLEG): container finished" podID="6ed1f144-c0f4-470e-9bc2-fc4ea36bc158" containerID="03ea6d6e185a67b99344dfb4cc4d82e9e0d863bc76b700c811e021763a985cb7" exitCode=0 Mar 21 08:41:41 crc kubenswrapper[4696]: I0321 08:41:41.625754 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9" event={"ID":"6ed1f144-c0f4-470e-9bc2-fc4ea36bc158","Type":"ContainerDied","Data":"03ea6d6e185a67b99344dfb4cc4d82e9e0d863bc76b700c811e021763a985cb7"} Mar 21 08:41:42 crc kubenswrapper[4696]: I0321 08:41:42.542480 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1d05966-d105-4c45-bfa2-e66a67882baf" path="/var/lib/kubelet/pods/d1d05966-d105-4c45-bfa2-e66a67882baf/volumes" Mar 21 08:41:42 crc kubenswrapper[4696]: I0321 08:41:42.638234 4696 generic.go:334] "Generic (PLEG): container finished" podID="6ed1f144-c0f4-470e-9bc2-fc4ea36bc158" containerID="568f892f2dee5e0285b08e8747bd09d0766e15f818e0e966af3d9346cbf04519" exitCode=0 Mar 21 08:41:42 crc kubenswrapper[4696]: I0321 08:41:42.638282 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9" event={"ID":"6ed1f144-c0f4-470e-9bc2-fc4ea36bc158","Type":"ContainerDied","Data":"568f892f2dee5e0285b08e8747bd09d0766e15f818e0e966af3d9346cbf04519"} Mar 21 08:41:43 crc kubenswrapper[4696]: I0321 08:41:43.931327 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9" Mar 21 08:41:44 crc kubenswrapper[4696]: I0321 08:41:44.046334 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6ed1f144-c0f4-470e-9bc2-fc4ea36bc158-util\") pod \"6ed1f144-c0f4-470e-9bc2-fc4ea36bc158\" (UID: \"6ed1f144-c0f4-470e-9bc2-fc4ea36bc158\") " Mar 21 08:41:44 crc kubenswrapper[4696]: I0321 08:41:44.046410 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6ed1f144-c0f4-470e-9bc2-fc4ea36bc158-bundle\") pod \"6ed1f144-c0f4-470e-9bc2-fc4ea36bc158\" (UID: \"6ed1f144-c0f4-470e-9bc2-fc4ea36bc158\") " Mar 21 08:41:44 crc kubenswrapper[4696]: I0321 08:41:44.046498 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnkv2\" (UniqueName: \"kubernetes.io/projected/6ed1f144-c0f4-470e-9bc2-fc4ea36bc158-kube-api-access-lnkv2\") pod \"6ed1f144-c0f4-470e-9bc2-fc4ea36bc158\" (UID: \"6ed1f144-c0f4-470e-9bc2-fc4ea36bc158\") " Mar 21 08:41:44 crc kubenswrapper[4696]: I0321 08:41:44.047300 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ed1f144-c0f4-470e-9bc2-fc4ea36bc158-bundle" (OuterVolumeSpecName: "bundle") pod "6ed1f144-c0f4-470e-9bc2-fc4ea36bc158" (UID: "6ed1f144-c0f4-470e-9bc2-fc4ea36bc158"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:41:44 crc kubenswrapper[4696]: I0321 08:41:44.055982 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ed1f144-c0f4-470e-9bc2-fc4ea36bc158-kube-api-access-lnkv2" (OuterVolumeSpecName: "kube-api-access-lnkv2") pod "6ed1f144-c0f4-470e-9bc2-fc4ea36bc158" (UID: "6ed1f144-c0f4-470e-9bc2-fc4ea36bc158"). InnerVolumeSpecName "kube-api-access-lnkv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:41:44 crc kubenswrapper[4696]: I0321 08:41:44.060175 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ed1f144-c0f4-470e-9bc2-fc4ea36bc158-util" (OuterVolumeSpecName: "util") pod "6ed1f144-c0f4-470e-9bc2-fc4ea36bc158" (UID: "6ed1f144-c0f4-470e-9bc2-fc4ea36bc158"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:41:44 crc kubenswrapper[4696]: I0321 08:41:44.148143 4696 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6ed1f144-c0f4-470e-9bc2-fc4ea36bc158-util\") on node \"crc\" DevicePath \"\"" Mar 21 08:41:44 crc kubenswrapper[4696]: I0321 08:41:44.148181 4696 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6ed1f144-c0f4-470e-9bc2-fc4ea36bc158-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:41:44 crc kubenswrapper[4696]: I0321 08:41:44.148190 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnkv2\" (UniqueName: \"kubernetes.io/projected/6ed1f144-c0f4-470e-9bc2-fc4ea36bc158-kube-api-access-lnkv2\") on node \"crc\" DevicePath \"\"" Mar 21 08:41:44 crc kubenswrapper[4696]: I0321 08:41:44.651006 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9" event={"ID":"6ed1f144-c0f4-470e-9bc2-fc4ea36bc158","Type":"ContainerDied","Data":"de7359c4bb78fcd602b953a1f77681509968a6c23eccb438528c6b18eba0caa4"} Mar 21 08:41:44 crc kubenswrapper[4696]: I0321 08:41:44.651050 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de7359c4bb78fcd602b953a1f77681509968a6c23eccb438528c6b18eba0caa4" Mar 21 08:41:44 crc kubenswrapper[4696]: I0321 08:41:44.651142 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.561171 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-78876797bd-qsnl8"] Mar 21 08:41:54 crc kubenswrapper[4696]: E0321 08:41:54.561894 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ed1f144-c0f4-470e-9bc2-fc4ea36bc158" containerName="extract" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.561906 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ed1f144-c0f4-470e-9bc2-fc4ea36bc158" containerName="extract" Mar 21 08:41:54 crc kubenswrapper[4696]: E0321 08:41:54.561918 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ed1f144-c0f4-470e-9bc2-fc4ea36bc158" containerName="pull" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.561924 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ed1f144-c0f4-470e-9bc2-fc4ea36bc158" containerName="pull" Mar 21 08:41:54 crc kubenswrapper[4696]: E0321 08:41:54.561939 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d05966-d105-4c45-bfa2-e66a67882baf" containerName="console" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.561945 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d05966-d105-4c45-bfa2-e66a67882baf" containerName="console" Mar 21 08:41:54 crc kubenswrapper[4696]: E0321 08:41:54.561956 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ed1f144-c0f4-470e-9bc2-fc4ea36bc158" containerName="util" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.561961 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ed1f144-c0f4-470e-9bc2-fc4ea36bc158" containerName="util" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.562059 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ed1f144-c0f4-470e-9bc2-fc4ea36bc158" containerName="extract" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.562072 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1d05966-d105-4c45-bfa2-e66a67882baf" containerName="console" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.562481 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-78876797bd-qsnl8" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.564092 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-qbpv6" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.564332 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.570910 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.570987 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.570987 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.582560 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-78876797bd-qsnl8"] Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.705612 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qxzl\" (UniqueName: \"kubernetes.io/projected/5fe21876-0bc8-489f-b862-6d48be898f5b-kube-api-access-5qxzl\") pod \"metallb-operator-controller-manager-78876797bd-qsnl8\" (UID: \"5fe21876-0bc8-489f-b862-6d48be898f5b\") " pod="metallb-system/metallb-operator-controller-manager-78876797bd-qsnl8" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.705971 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5fe21876-0bc8-489f-b862-6d48be898f5b-webhook-cert\") pod \"metallb-operator-controller-manager-78876797bd-qsnl8\" (UID: \"5fe21876-0bc8-489f-b862-6d48be898f5b\") " pod="metallb-system/metallb-operator-controller-manager-78876797bd-qsnl8" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.706010 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5fe21876-0bc8-489f-b862-6d48be898f5b-apiservice-cert\") pod \"metallb-operator-controller-manager-78876797bd-qsnl8\" (UID: \"5fe21876-0bc8-489f-b862-6d48be898f5b\") " pod="metallb-system/metallb-operator-controller-manager-78876797bd-qsnl8" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.807042 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5fe21876-0bc8-489f-b862-6d48be898f5b-apiservice-cert\") pod \"metallb-operator-controller-manager-78876797bd-qsnl8\" (UID: \"5fe21876-0bc8-489f-b862-6d48be898f5b\") " pod="metallb-system/metallb-operator-controller-manager-78876797bd-qsnl8" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.807131 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qxzl\" (UniqueName: \"kubernetes.io/projected/5fe21876-0bc8-489f-b862-6d48be898f5b-kube-api-access-5qxzl\") pod \"metallb-operator-controller-manager-78876797bd-qsnl8\" (UID: \"5fe21876-0bc8-489f-b862-6d48be898f5b\") " pod="metallb-system/metallb-operator-controller-manager-78876797bd-qsnl8" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.807214 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5fe21876-0bc8-489f-b862-6d48be898f5b-webhook-cert\") pod \"metallb-operator-controller-manager-78876797bd-qsnl8\" (UID: \"5fe21876-0bc8-489f-b862-6d48be898f5b\") " pod="metallb-system/metallb-operator-controller-manager-78876797bd-qsnl8" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.812969 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5fe21876-0bc8-489f-b862-6d48be898f5b-apiservice-cert\") pod \"metallb-operator-controller-manager-78876797bd-qsnl8\" (UID: \"5fe21876-0bc8-489f-b862-6d48be898f5b\") " pod="metallb-system/metallb-operator-controller-manager-78876797bd-qsnl8" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.813164 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5fe21876-0bc8-489f-b862-6d48be898f5b-webhook-cert\") pod \"metallb-operator-controller-manager-78876797bd-qsnl8\" (UID: \"5fe21876-0bc8-489f-b862-6d48be898f5b\") " pod="metallb-system/metallb-operator-controller-manager-78876797bd-qsnl8" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.839561 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qxzl\" (UniqueName: \"kubernetes.io/projected/5fe21876-0bc8-489f-b862-6d48be898f5b-kube-api-access-5qxzl\") pod \"metallb-operator-controller-manager-78876797bd-qsnl8\" (UID: \"5fe21876-0bc8-489f-b862-6d48be898f5b\") " pod="metallb-system/metallb-operator-controller-manager-78876797bd-qsnl8" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.861596 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7749f78475-cjlsc"] Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.862415 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7749f78475-cjlsc" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.875787 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.875991 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-2fcr9" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.876125 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.880663 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7749f78475-cjlsc"] Mar 21 08:41:54 crc kubenswrapper[4696]: I0321 08:41:54.934096 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-78876797bd-qsnl8" Mar 21 08:41:55 crc kubenswrapper[4696]: I0321 08:41:55.009067 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkgbp\" (UniqueName: \"kubernetes.io/projected/dd16e737-b1c3-451b-a87c-90ec18d32dd3-kube-api-access-bkgbp\") pod \"metallb-operator-webhook-server-7749f78475-cjlsc\" (UID: \"dd16e737-b1c3-451b-a87c-90ec18d32dd3\") " pod="metallb-system/metallb-operator-webhook-server-7749f78475-cjlsc" Mar 21 08:41:55 crc kubenswrapper[4696]: I0321 08:41:55.009329 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dd16e737-b1c3-451b-a87c-90ec18d32dd3-apiservice-cert\") pod \"metallb-operator-webhook-server-7749f78475-cjlsc\" (UID: \"dd16e737-b1c3-451b-a87c-90ec18d32dd3\") " pod="metallb-system/metallb-operator-webhook-server-7749f78475-cjlsc" Mar 21 08:41:55 crc kubenswrapper[4696]: I0321 08:41:55.009384 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dd16e737-b1c3-451b-a87c-90ec18d32dd3-webhook-cert\") pod \"metallb-operator-webhook-server-7749f78475-cjlsc\" (UID: \"dd16e737-b1c3-451b-a87c-90ec18d32dd3\") " pod="metallb-system/metallb-operator-webhook-server-7749f78475-cjlsc" Mar 21 08:41:55 crc kubenswrapper[4696]: I0321 08:41:55.113982 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dd16e737-b1c3-451b-a87c-90ec18d32dd3-apiservice-cert\") pod \"metallb-operator-webhook-server-7749f78475-cjlsc\" (UID: \"dd16e737-b1c3-451b-a87c-90ec18d32dd3\") " pod="metallb-system/metallb-operator-webhook-server-7749f78475-cjlsc" Mar 21 08:41:55 crc kubenswrapper[4696]: I0321 08:41:55.114051 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dd16e737-b1c3-451b-a87c-90ec18d32dd3-webhook-cert\") pod \"metallb-operator-webhook-server-7749f78475-cjlsc\" (UID: \"dd16e737-b1c3-451b-a87c-90ec18d32dd3\") " pod="metallb-system/metallb-operator-webhook-server-7749f78475-cjlsc" Mar 21 08:41:55 crc kubenswrapper[4696]: I0321 08:41:55.114073 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkgbp\" (UniqueName: \"kubernetes.io/projected/dd16e737-b1c3-451b-a87c-90ec18d32dd3-kube-api-access-bkgbp\") pod \"metallb-operator-webhook-server-7749f78475-cjlsc\" (UID: \"dd16e737-b1c3-451b-a87c-90ec18d32dd3\") " pod="metallb-system/metallb-operator-webhook-server-7749f78475-cjlsc" Mar 21 08:41:55 crc kubenswrapper[4696]: I0321 08:41:55.119442 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dd16e737-b1c3-451b-a87c-90ec18d32dd3-apiservice-cert\") pod \"metallb-operator-webhook-server-7749f78475-cjlsc\" (UID: \"dd16e737-b1c3-451b-a87c-90ec18d32dd3\") " pod="metallb-system/metallb-operator-webhook-server-7749f78475-cjlsc" Mar 21 08:41:55 crc kubenswrapper[4696]: I0321 08:41:55.119475 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dd16e737-b1c3-451b-a87c-90ec18d32dd3-webhook-cert\") pod \"metallb-operator-webhook-server-7749f78475-cjlsc\" (UID: \"dd16e737-b1c3-451b-a87c-90ec18d32dd3\") " pod="metallb-system/metallb-operator-webhook-server-7749f78475-cjlsc" Mar 21 08:41:55 crc kubenswrapper[4696]: I0321 08:41:55.130794 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkgbp\" (UniqueName: \"kubernetes.io/projected/dd16e737-b1c3-451b-a87c-90ec18d32dd3-kube-api-access-bkgbp\") pod \"metallb-operator-webhook-server-7749f78475-cjlsc\" (UID: \"dd16e737-b1c3-451b-a87c-90ec18d32dd3\") " pod="metallb-system/metallb-operator-webhook-server-7749f78475-cjlsc" Mar 21 08:41:55 crc kubenswrapper[4696]: I0321 08:41:55.194564 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7749f78475-cjlsc" Mar 21 08:41:55 crc kubenswrapper[4696]: I0321 08:41:55.397381 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-78876797bd-qsnl8"] Mar 21 08:41:55 crc kubenswrapper[4696]: I0321 08:41:55.423656 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7749f78475-cjlsc"] Mar 21 08:41:55 crc kubenswrapper[4696]: W0321 08:41:55.433373 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd16e737_b1c3_451b_a87c_90ec18d32dd3.slice/crio-ea3ca5158390f9c510b8ebc7ddff23a64bc1488e91c3dcfea829a375ffae26b5 WatchSource:0}: Error finding container ea3ca5158390f9c510b8ebc7ddff23a64bc1488e91c3dcfea829a375ffae26b5: Status 404 returned error can't find the container with id ea3ca5158390f9c510b8ebc7ddff23a64bc1488e91c3dcfea829a375ffae26b5 Mar 21 08:41:55 crc kubenswrapper[4696]: I0321 08:41:55.715630 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-78876797bd-qsnl8" event={"ID":"5fe21876-0bc8-489f-b862-6d48be898f5b","Type":"ContainerStarted","Data":"30b644ea3000f8605673836308740b809dff1c9786821e12b268bd4ba621e407"} Mar 21 08:41:55 crc kubenswrapper[4696]: I0321 08:41:55.716699 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7749f78475-cjlsc" event={"ID":"dd16e737-b1c3-451b-a87c-90ec18d32dd3","Type":"ContainerStarted","Data":"ea3ca5158390f9c510b8ebc7ddff23a64bc1488e91c3dcfea829a375ffae26b5"} Mar 21 08:41:59 crc kubenswrapper[4696]: I0321 08:41:59.753175 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-78876797bd-qsnl8" event={"ID":"5fe21876-0bc8-489f-b862-6d48be898f5b","Type":"ContainerStarted","Data":"ea493be84b7a72a4a3e1961294efda51855ffd9127291f088e2c5a1d0670e94c"} Mar 21 08:41:59 crc kubenswrapper[4696]: I0321 08:41:59.753678 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-78876797bd-qsnl8" Mar 21 08:42:00 crc kubenswrapper[4696]: I0321 08:42:00.127971 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-78876797bd-qsnl8" podStartSLOduration=2.537280944 podStartE2EDuration="6.127952738s" podCreationTimestamp="2026-03-21 08:41:54 +0000 UTC" firstStartedPulling="2026-03-21 08:41:55.415544805 +0000 UTC m=+849.536425518" lastFinishedPulling="2026-03-21 08:41:59.006216599 +0000 UTC m=+853.127097312" observedRunningTime="2026-03-21 08:41:59.783761294 +0000 UTC m=+853.904642027" watchObservedRunningTime="2026-03-21 08:42:00.127952738 +0000 UTC m=+854.248833451" Mar 21 08:42:00 crc kubenswrapper[4696]: I0321 08:42:00.138753 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568042-xb457"] Mar 21 08:42:00 crc kubenswrapper[4696]: I0321 08:42:00.140050 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568042-xb457"] Mar 21 08:42:00 crc kubenswrapper[4696]: I0321 08:42:00.140140 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568042-xb457" Mar 21 08:42:00 crc kubenswrapper[4696]: I0321 08:42:00.172909 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 08:42:00 crc kubenswrapper[4696]: I0321 08:42:00.173225 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 08:42:00 crc kubenswrapper[4696]: I0321 08:42:00.173519 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 08:42:00 crc kubenswrapper[4696]: I0321 08:42:00.184683 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82xzj\" (UniqueName: \"kubernetes.io/projected/235bf627-f528-4aa9-85c1-b069ac61f695-kube-api-access-82xzj\") pod \"auto-csr-approver-29568042-xb457\" (UID: \"235bf627-f528-4aa9-85c1-b069ac61f695\") " pod="openshift-infra/auto-csr-approver-29568042-xb457" Mar 21 08:42:00 crc kubenswrapper[4696]: I0321 08:42:00.285789 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82xzj\" (UniqueName: \"kubernetes.io/projected/235bf627-f528-4aa9-85c1-b069ac61f695-kube-api-access-82xzj\") pod \"auto-csr-approver-29568042-xb457\" (UID: \"235bf627-f528-4aa9-85c1-b069ac61f695\") " pod="openshift-infra/auto-csr-approver-29568042-xb457" Mar 21 08:42:00 crc kubenswrapper[4696]: I0321 08:42:00.306151 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82xzj\" (UniqueName: \"kubernetes.io/projected/235bf627-f528-4aa9-85c1-b069ac61f695-kube-api-access-82xzj\") pod \"auto-csr-approver-29568042-xb457\" (UID: \"235bf627-f528-4aa9-85c1-b069ac61f695\") " pod="openshift-infra/auto-csr-approver-29568042-xb457" Mar 21 08:42:00 crc kubenswrapper[4696]: I0321 08:42:00.493391 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568042-xb457" Mar 21 08:42:00 crc kubenswrapper[4696]: I0321 08:42:00.766196 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7749f78475-cjlsc" event={"ID":"dd16e737-b1c3-451b-a87c-90ec18d32dd3","Type":"ContainerStarted","Data":"c70215017fc5c068dec588ca7090e2635584e417819898852e0848d1743a1347"} Mar 21 08:42:00 crc kubenswrapper[4696]: I0321 08:42:00.786773 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7749f78475-cjlsc" podStartSLOduration=1.871905197 podStartE2EDuration="6.786755737s" podCreationTimestamp="2026-03-21 08:41:54 +0000 UTC" firstStartedPulling="2026-03-21 08:41:55.436804269 +0000 UTC m=+849.557684982" lastFinishedPulling="2026-03-21 08:42:00.351654809 +0000 UTC m=+854.472535522" observedRunningTime="2026-03-21 08:42:00.786353307 +0000 UTC m=+854.907234040" watchObservedRunningTime="2026-03-21 08:42:00.786755737 +0000 UTC m=+854.907636450" Mar 21 08:42:00 crc kubenswrapper[4696]: I0321 08:42:00.899737 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568042-xb457"] Mar 21 08:42:01 crc kubenswrapper[4696]: I0321 08:42:01.782366 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568042-xb457" event={"ID":"235bf627-f528-4aa9-85c1-b069ac61f695","Type":"ContainerStarted","Data":"1e828c51598c56cfe0a55ae307a4c0922c3c54dd571f3a57350d120177dfa3a5"} Mar 21 08:42:01 crc kubenswrapper[4696]: I0321 08:42:01.782492 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7749f78475-cjlsc" Mar 21 08:42:02 crc kubenswrapper[4696]: I0321 08:42:02.793069 4696 generic.go:334] "Generic (PLEG): container finished" podID="235bf627-f528-4aa9-85c1-b069ac61f695" containerID="fe52e941866d9eb50ec7713e66e65174156613e23766607e0802e44303d6d7f6" exitCode=0 Mar 21 08:42:02 crc kubenswrapper[4696]: I0321 08:42:02.793142 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568042-xb457" event={"ID":"235bf627-f528-4aa9-85c1-b069ac61f695","Type":"ContainerDied","Data":"fe52e941866d9eb50ec7713e66e65174156613e23766607e0802e44303d6d7f6"} Mar 21 08:42:04 crc kubenswrapper[4696]: I0321 08:42:04.085682 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568042-xb457" Mar 21 08:42:04 crc kubenswrapper[4696]: I0321 08:42:04.158237 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82xzj\" (UniqueName: \"kubernetes.io/projected/235bf627-f528-4aa9-85c1-b069ac61f695-kube-api-access-82xzj\") pod \"235bf627-f528-4aa9-85c1-b069ac61f695\" (UID: \"235bf627-f528-4aa9-85c1-b069ac61f695\") " Mar 21 08:42:04 crc kubenswrapper[4696]: I0321 08:42:04.163580 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/235bf627-f528-4aa9-85c1-b069ac61f695-kube-api-access-82xzj" (OuterVolumeSpecName: "kube-api-access-82xzj") pod "235bf627-f528-4aa9-85c1-b069ac61f695" (UID: "235bf627-f528-4aa9-85c1-b069ac61f695"). InnerVolumeSpecName "kube-api-access-82xzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:42:04 crc kubenswrapper[4696]: I0321 08:42:04.259938 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82xzj\" (UniqueName: \"kubernetes.io/projected/235bf627-f528-4aa9-85c1-b069ac61f695-kube-api-access-82xzj\") on node \"crc\" DevicePath \"\"" Mar 21 08:42:04 crc kubenswrapper[4696]: I0321 08:42:04.804655 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568042-xb457" event={"ID":"235bf627-f528-4aa9-85c1-b069ac61f695","Type":"ContainerDied","Data":"1e828c51598c56cfe0a55ae307a4c0922c3c54dd571f3a57350d120177dfa3a5"} Mar 21 08:42:04 crc kubenswrapper[4696]: I0321 08:42:04.804693 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e828c51598c56cfe0a55ae307a4c0922c3c54dd571f3a57350d120177dfa3a5" Mar 21 08:42:04 crc kubenswrapper[4696]: I0321 08:42:04.804742 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568042-xb457" Mar 21 08:42:05 crc kubenswrapper[4696]: I0321 08:42:05.127334 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568036-lv7tq"] Mar 21 08:42:05 crc kubenswrapper[4696]: I0321 08:42:05.134156 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568036-lv7tq"] Mar 21 08:42:06 crc kubenswrapper[4696]: I0321 08:42:06.544743 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0cac1ea-786f-41a3-8a4c-0d8d02b40b86" path="/var/lib/kubelet/pods/d0cac1ea-786f-41a3-8a4c-0d8d02b40b86/volumes" Mar 21 08:42:15 crc kubenswrapper[4696]: I0321 08:42:15.201006 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7749f78475-cjlsc" Mar 21 08:42:15 crc kubenswrapper[4696]: I0321 08:42:15.691515 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rx6rk"] Mar 21 08:42:15 crc kubenswrapper[4696]: E0321 08:42:15.692108 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="235bf627-f528-4aa9-85c1-b069ac61f695" containerName="oc" Mar 21 08:42:15 crc kubenswrapper[4696]: I0321 08:42:15.692130 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="235bf627-f528-4aa9-85c1-b069ac61f695" containerName="oc" Mar 21 08:42:15 crc kubenswrapper[4696]: I0321 08:42:15.692252 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="235bf627-f528-4aa9-85c1-b069ac61f695" containerName="oc" Mar 21 08:42:15 crc kubenswrapper[4696]: I0321 08:42:15.693195 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rx6rk" Mar 21 08:42:15 crc kubenswrapper[4696]: I0321 08:42:15.695088 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m4sj\" (UniqueName: \"kubernetes.io/projected/463ad4fb-ee92-4eee-99a9-57066f7706c6-kube-api-access-6m4sj\") pod \"community-operators-rx6rk\" (UID: \"463ad4fb-ee92-4eee-99a9-57066f7706c6\") " pod="openshift-marketplace/community-operators-rx6rk" Mar 21 08:42:15 crc kubenswrapper[4696]: I0321 08:42:15.695242 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/463ad4fb-ee92-4eee-99a9-57066f7706c6-catalog-content\") pod \"community-operators-rx6rk\" (UID: \"463ad4fb-ee92-4eee-99a9-57066f7706c6\") " pod="openshift-marketplace/community-operators-rx6rk" Mar 21 08:42:15 crc kubenswrapper[4696]: I0321 08:42:15.695299 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/463ad4fb-ee92-4eee-99a9-57066f7706c6-utilities\") pod \"community-operators-rx6rk\" (UID: \"463ad4fb-ee92-4eee-99a9-57066f7706c6\") " pod="openshift-marketplace/community-operators-rx6rk" Mar 21 08:42:15 crc kubenswrapper[4696]: I0321 08:42:15.699223 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rx6rk"] Mar 21 08:42:15 crc kubenswrapper[4696]: I0321 08:42:15.796014 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/463ad4fb-ee92-4eee-99a9-57066f7706c6-catalog-content\") pod \"community-operators-rx6rk\" (UID: \"463ad4fb-ee92-4eee-99a9-57066f7706c6\") " pod="openshift-marketplace/community-operators-rx6rk" Mar 21 08:42:15 crc kubenswrapper[4696]: I0321 08:42:15.796069 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/463ad4fb-ee92-4eee-99a9-57066f7706c6-utilities\") pod \"community-operators-rx6rk\" (UID: \"463ad4fb-ee92-4eee-99a9-57066f7706c6\") " pod="openshift-marketplace/community-operators-rx6rk" Mar 21 08:42:15 crc kubenswrapper[4696]: I0321 08:42:15.796098 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m4sj\" (UniqueName: \"kubernetes.io/projected/463ad4fb-ee92-4eee-99a9-57066f7706c6-kube-api-access-6m4sj\") pod \"community-operators-rx6rk\" (UID: \"463ad4fb-ee92-4eee-99a9-57066f7706c6\") " pod="openshift-marketplace/community-operators-rx6rk" Mar 21 08:42:15 crc kubenswrapper[4696]: I0321 08:42:15.796524 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/463ad4fb-ee92-4eee-99a9-57066f7706c6-utilities\") pod \"community-operators-rx6rk\" (UID: \"463ad4fb-ee92-4eee-99a9-57066f7706c6\") " pod="openshift-marketplace/community-operators-rx6rk" Mar 21 08:42:15 crc kubenswrapper[4696]: I0321 08:42:15.796538 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/463ad4fb-ee92-4eee-99a9-57066f7706c6-catalog-content\") pod \"community-operators-rx6rk\" (UID: \"463ad4fb-ee92-4eee-99a9-57066f7706c6\") " pod="openshift-marketplace/community-operators-rx6rk" Mar 21 08:42:15 crc kubenswrapper[4696]: I0321 08:42:15.824549 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m4sj\" (UniqueName: \"kubernetes.io/projected/463ad4fb-ee92-4eee-99a9-57066f7706c6-kube-api-access-6m4sj\") pod \"community-operators-rx6rk\" (UID: \"463ad4fb-ee92-4eee-99a9-57066f7706c6\") " pod="openshift-marketplace/community-operators-rx6rk" Mar 21 08:42:16 crc kubenswrapper[4696]: I0321 08:42:16.011287 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rx6rk" Mar 21 08:42:16 crc kubenswrapper[4696]: I0321 08:42:16.463944 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rx6rk"] Mar 21 08:42:16 crc kubenswrapper[4696]: W0321 08:42:16.471078 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod463ad4fb_ee92_4eee_99a9_57066f7706c6.slice/crio-0455e50b514ecbb1774976038e73fb703d175e658c8b1bfc9e229d49fce0a582 WatchSource:0}: Error finding container 0455e50b514ecbb1774976038e73fb703d175e658c8b1bfc9e229d49fce0a582: Status 404 returned error can't find the container with id 0455e50b514ecbb1774976038e73fb703d175e658c8b1bfc9e229d49fce0a582 Mar 21 08:42:16 crc kubenswrapper[4696]: I0321 08:42:16.873698 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rx6rk" event={"ID":"463ad4fb-ee92-4eee-99a9-57066f7706c6","Type":"ContainerStarted","Data":"0455e50b514ecbb1774976038e73fb703d175e658c8b1bfc9e229d49fce0a582"} Mar 21 08:42:18 crc kubenswrapper[4696]: I0321 08:42:18.886233 4696 generic.go:334] "Generic (PLEG): container finished" podID="463ad4fb-ee92-4eee-99a9-57066f7706c6" containerID="19c556e08c02cb3f061d1f13826124dfa3e78075b72c45c534667497ad813cf8" exitCode=0 Mar 21 08:42:18 crc kubenswrapper[4696]: I0321 08:42:18.886337 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rx6rk" event={"ID":"463ad4fb-ee92-4eee-99a9-57066f7706c6","Type":"ContainerDied","Data":"19c556e08c02cb3f061d1f13826124dfa3e78075b72c45c534667497ad813cf8"} Mar 21 08:42:19 crc kubenswrapper[4696]: I0321 08:42:19.895567 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rx6rk" event={"ID":"463ad4fb-ee92-4eee-99a9-57066f7706c6","Type":"ContainerStarted","Data":"da9903058911bb047188d07421f4936cfc4703784d9959c576a870d3901c4a1a"} Mar 21 08:42:20 crc kubenswrapper[4696]: I0321 08:42:20.903472 4696 generic.go:334] "Generic (PLEG): container finished" podID="463ad4fb-ee92-4eee-99a9-57066f7706c6" containerID="da9903058911bb047188d07421f4936cfc4703784d9959c576a870d3901c4a1a" exitCode=0 Mar 21 08:42:20 crc kubenswrapper[4696]: I0321 08:42:20.903566 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rx6rk" event={"ID":"463ad4fb-ee92-4eee-99a9-57066f7706c6","Type":"ContainerDied","Data":"da9903058911bb047188d07421f4936cfc4703784d9959c576a870d3901c4a1a"} Mar 21 08:42:21 crc kubenswrapper[4696]: I0321 08:42:21.911624 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rx6rk" event={"ID":"463ad4fb-ee92-4eee-99a9-57066f7706c6","Type":"ContainerStarted","Data":"347f5c0d36fd328f2dd9fceb3a288030acb9adebabf594d4a5f4063f86f1c102"} Mar 21 08:42:21 crc kubenswrapper[4696]: I0321 08:42:21.929053 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rx6rk" podStartSLOduration=4.5039021649999995 podStartE2EDuration="6.929026927s" podCreationTimestamp="2026-03-21 08:42:15 +0000 UTC" firstStartedPulling="2026-03-21 08:42:18.888538258 +0000 UTC m=+873.009418981" lastFinishedPulling="2026-03-21 08:42:21.313663 +0000 UTC m=+875.434543743" observedRunningTime="2026-03-21 08:42:21.924417483 +0000 UTC m=+876.045298196" watchObservedRunningTime="2026-03-21 08:42:21.929026927 +0000 UTC m=+876.049907670" Mar 21 08:42:26 crc kubenswrapper[4696]: I0321 08:42:26.012030 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rx6rk" Mar 21 08:42:26 crc kubenswrapper[4696]: I0321 08:42:26.012590 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rx6rk" Mar 21 08:42:26 crc kubenswrapper[4696]: I0321 08:42:26.052318 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rx6rk" Mar 21 08:42:26 crc kubenswrapper[4696]: I0321 08:42:26.981148 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rx6rk" Mar 21 08:42:27 crc kubenswrapper[4696]: I0321 08:42:27.684561 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rx6rk"] Mar 21 08:42:28 crc kubenswrapper[4696]: I0321 08:42:28.950527 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rx6rk" podUID="463ad4fb-ee92-4eee-99a9-57066f7706c6" containerName="registry-server" containerID="cri-o://347f5c0d36fd328f2dd9fceb3a288030acb9adebabf594d4a5f4063f86f1c102" gracePeriod=2 Mar 21 08:42:29 crc kubenswrapper[4696]: I0321 08:42:29.301707 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rx6rk" Mar 21 08:42:29 crc kubenswrapper[4696]: I0321 08:42:29.467414 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/463ad4fb-ee92-4eee-99a9-57066f7706c6-utilities\") pod \"463ad4fb-ee92-4eee-99a9-57066f7706c6\" (UID: \"463ad4fb-ee92-4eee-99a9-57066f7706c6\") " Mar 21 08:42:29 crc kubenswrapper[4696]: I0321 08:42:29.467559 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6m4sj\" (UniqueName: \"kubernetes.io/projected/463ad4fb-ee92-4eee-99a9-57066f7706c6-kube-api-access-6m4sj\") pod \"463ad4fb-ee92-4eee-99a9-57066f7706c6\" (UID: \"463ad4fb-ee92-4eee-99a9-57066f7706c6\") " Mar 21 08:42:29 crc kubenswrapper[4696]: I0321 08:42:29.467595 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/463ad4fb-ee92-4eee-99a9-57066f7706c6-catalog-content\") pod \"463ad4fb-ee92-4eee-99a9-57066f7706c6\" (UID: \"463ad4fb-ee92-4eee-99a9-57066f7706c6\") " Mar 21 08:42:29 crc kubenswrapper[4696]: I0321 08:42:29.468316 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/463ad4fb-ee92-4eee-99a9-57066f7706c6-utilities" (OuterVolumeSpecName: "utilities") pod "463ad4fb-ee92-4eee-99a9-57066f7706c6" (UID: "463ad4fb-ee92-4eee-99a9-57066f7706c6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:42:29 crc kubenswrapper[4696]: I0321 08:42:29.472256 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/463ad4fb-ee92-4eee-99a9-57066f7706c6-kube-api-access-6m4sj" (OuterVolumeSpecName: "kube-api-access-6m4sj") pod "463ad4fb-ee92-4eee-99a9-57066f7706c6" (UID: "463ad4fb-ee92-4eee-99a9-57066f7706c6"). InnerVolumeSpecName "kube-api-access-6m4sj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:42:29 crc kubenswrapper[4696]: I0321 08:42:29.522620 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/463ad4fb-ee92-4eee-99a9-57066f7706c6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "463ad4fb-ee92-4eee-99a9-57066f7706c6" (UID: "463ad4fb-ee92-4eee-99a9-57066f7706c6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:42:29 crc kubenswrapper[4696]: I0321 08:42:29.568760 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6m4sj\" (UniqueName: \"kubernetes.io/projected/463ad4fb-ee92-4eee-99a9-57066f7706c6-kube-api-access-6m4sj\") on node \"crc\" DevicePath \"\"" Mar 21 08:42:29 crc kubenswrapper[4696]: I0321 08:42:29.568798 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/463ad4fb-ee92-4eee-99a9-57066f7706c6-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 08:42:29 crc kubenswrapper[4696]: I0321 08:42:29.568807 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/463ad4fb-ee92-4eee-99a9-57066f7706c6-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 08:42:29 crc kubenswrapper[4696]: I0321 08:42:29.959693 4696 generic.go:334] "Generic (PLEG): container finished" podID="463ad4fb-ee92-4eee-99a9-57066f7706c6" containerID="347f5c0d36fd328f2dd9fceb3a288030acb9adebabf594d4a5f4063f86f1c102" exitCode=0 Mar 21 08:42:29 crc kubenswrapper[4696]: I0321 08:42:29.959755 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rx6rk" event={"ID":"463ad4fb-ee92-4eee-99a9-57066f7706c6","Type":"ContainerDied","Data":"347f5c0d36fd328f2dd9fceb3a288030acb9adebabf594d4a5f4063f86f1c102"} Mar 21 08:42:29 crc kubenswrapper[4696]: I0321 08:42:29.959803 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rx6rk" Mar 21 08:42:29 crc kubenswrapper[4696]: I0321 08:42:29.959968 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rx6rk" event={"ID":"463ad4fb-ee92-4eee-99a9-57066f7706c6","Type":"ContainerDied","Data":"0455e50b514ecbb1774976038e73fb703d175e658c8b1bfc9e229d49fce0a582"} Mar 21 08:42:29 crc kubenswrapper[4696]: I0321 08:42:29.960007 4696 scope.go:117] "RemoveContainer" containerID="347f5c0d36fd328f2dd9fceb3a288030acb9adebabf594d4a5f4063f86f1c102" Mar 21 08:42:29 crc kubenswrapper[4696]: I0321 08:42:29.986045 4696 scope.go:117] "RemoveContainer" containerID="da9903058911bb047188d07421f4936cfc4703784d9959c576a870d3901c4a1a" Mar 21 08:42:30 crc kubenswrapper[4696]: I0321 08:42:30.012567 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rx6rk"] Mar 21 08:42:30 crc kubenswrapper[4696]: I0321 08:42:30.017764 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rx6rk"] Mar 21 08:42:30 crc kubenswrapper[4696]: I0321 08:42:30.025006 4696 scope.go:117] "RemoveContainer" containerID="19c556e08c02cb3f061d1f13826124dfa3e78075b72c45c534667497ad813cf8" Mar 21 08:42:30 crc kubenswrapper[4696]: I0321 08:42:30.043743 4696 scope.go:117] "RemoveContainer" containerID="347f5c0d36fd328f2dd9fceb3a288030acb9adebabf594d4a5f4063f86f1c102" Mar 21 08:42:30 crc kubenswrapper[4696]: E0321 08:42:30.044219 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"347f5c0d36fd328f2dd9fceb3a288030acb9adebabf594d4a5f4063f86f1c102\": container with ID starting with 347f5c0d36fd328f2dd9fceb3a288030acb9adebabf594d4a5f4063f86f1c102 not found: ID does not exist" containerID="347f5c0d36fd328f2dd9fceb3a288030acb9adebabf594d4a5f4063f86f1c102" Mar 21 08:42:30 crc kubenswrapper[4696]: I0321 08:42:30.044278 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"347f5c0d36fd328f2dd9fceb3a288030acb9adebabf594d4a5f4063f86f1c102"} err="failed to get container status \"347f5c0d36fd328f2dd9fceb3a288030acb9adebabf594d4a5f4063f86f1c102\": rpc error: code = NotFound desc = could not find container \"347f5c0d36fd328f2dd9fceb3a288030acb9adebabf594d4a5f4063f86f1c102\": container with ID starting with 347f5c0d36fd328f2dd9fceb3a288030acb9adebabf594d4a5f4063f86f1c102 not found: ID does not exist" Mar 21 08:42:30 crc kubenswrapper[4696]: I0321 08:42:30.044312 4696 scope.go:117] "RemoveContainer" containerID="da9903058911bb047188d07421f4936cfc4703784d9959c576a870d3901c4a1a" Mar 21 08:42:30 crc kubenswrapper[4696]: E0321 08:42:30.044962 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da9903058911bb047188d07421f4936cfc4703784d9959c576a870d3901c4a1a\": container with ID starting with da9903058911bb047188d07421f4936cfc4703784d9959c576a870d3901c4a1a not found: ID does not exist" containerID="da9903058911bb047188d07421f4936cfc4703784d9959c576a870d3901c4a1a" Mar 21 08:42:30 crc kubenswrapper[4696]: I0321 08:42:30.045004 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da9903058911bb047188d07421f4936cfc4703784d9959c576a870d3901c4a1a"} err="failed to get container status \"da9903058911bb047188d07421f4936cfc4703784d9959c576a870d3901c4a1a\": rpc error: code = NotFound desc = could not find container \"da9903058911bb047188d07421f4936cfc4703784d9959c576a870d3901c4a1a\": container with ID starting with da9903058911bb047188d07421f4936cfc4703784d9959c576a870d3901c4a1a not found: ID does not exist" Mar 21 08:42:30 crc kubenswrapper[4696]: I0321 08:42:30.045034 4696 scope.go:117] "RemoveContainer" containerID="19c556e08c02cb3f061d1f13826124dfa3e78075b72c45c534667497ad813cf8" Mar 21 08:42:30 crc kubenswrapper[4696]: E0321 08:42:30.045289 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19c556e08c02cb3f061d1f13826124dfa3e78075b72c45c534667497ad813cf8\": container with ID starting with 19c556e08c02cb3f061d1f13826124dfa3e78075b72c45c534667497ad813cf8 not found: ID does not exist" containerID="19c556e08c02cb3f061d1f13826124dfa3e78075b72c45c534667497ad813cf8" Mar 21 08:42:30 crc kubenswrapper[4696]: I0321 08:42:30.045320 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19c556e08c02cb3f061d1f13826124dfa3e78075b72c45c534667497ad813cf8"} err="failed to get container status \"19c556e08c02cb3f061d1f13826124dfa3e78075b72c45c534667497ad813cf8\": rpc error: code = NotFound desc = could not find container \"19c556e08c02cb3f061d1f13826124dfa3e78075b72c45c534667497ad813cf8\": container with ID starting with 19c556e08c02cb3f061d1f13826124dfa3e78075b72c45c534667497ad813cf8 not found: ID does not exist" Mar 21 08:42:30 crc kubenswrapper[4696]: I0321 08:42:30.545974 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="463ad4fb-ee92-4eee-99a9-57066f7706c6" path="/var/lib/kubelet/pods/463ad4fb-ee92-4eee-99a9-57066f7706c6/volumes" Mar 21 08:42:34 crc kubenswrapper[4696]: I0321 08:42:34.938394 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-78876797bd-qsnl8" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.610540 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-7v4k6"] Mar 21 08:42:35 crc kubenswrapper[4696]: E0321 08:42:35.611114 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="463ad4fb-ee92-4eee-99a9-57066f7706c6" containerName="registry-server" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.611130 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="463ad4fb-ee92-4eee-99a9-57066f7706c6" containerName="registry-server" Mar 21 08:42:35 crc kubenswrapper[4696]: E0321 08:42:35.611146 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="463ad4fb-ee92-4eee-99a9-57066f7706c6" containerName="extract-content" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.611154 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="463ad4fb-ee92-4eee-99a9-57066f7706c6" containerName="extract-content" Mar 21 08:42:35 crc kubenswrapper[4696]: E0321 08:42:35.611163 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="463ad4fb-ee92-4eee-99a9-57066f7706c6" containerName="extract-utilities" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.611169 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="463ad4fb-ee92-4eee-99a9-57066f7706c6" containerName="extract-utilities" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.611278 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="463ad4fb-ee92-4eee-99a9-57066f7706c6" containerName="registry-server" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.613351 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.614924 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-gk8sz"] Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.615462 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-r547c" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.615521 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.615682 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-gk8sz" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.617213 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.618696 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.641384 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-gk8sz"] Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.700325 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-zfggl"] Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.701372 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-zfggl" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.703311 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.703325 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.704288 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.704614 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-km8jm" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.716654 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-7bb4cc7c98-lzh5g"] Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.717583 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-lzh5g" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.723995 4696 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.730132 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-lzh5g"] Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.744150 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg7qh\" (UniqueName: \"kubernetes.io/projected/69b8f173-b1f8-45e3-a6ae-ffb987fc709e-kube-api-access-wg7qh\") pod \"frr-k8s-webhook-server-bcc4b6f68-gk8sz\" (UID: \"69b8f173-b1f8-45e3-a6ae-ffb987fc709e\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-gk8sz" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.744424 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4f243b13-1046-485b-b726-469b5a2b283f-metrics\") pod \"frr-k8s-7v4k6\" (UID: \"4f243b13-1046-485b-b726-469b5a2b283f\") " pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.744536 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4f243b13-1046-485b-b726-469b5a2b283f-metrics-certs\") pod \"frr-k8s-7v4k6\" (UID: \"4f243b13-1046-485b-b726-469b5a2b283f\") " pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.744650 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4f243b13-1046-485b-b726-469b5a2b283f-frr-sockets\") pod \"frr-k8s-7v4k6\" (UID: \"4f243b13-1046-485b-b726-469b5a2b283f\") " pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.744856 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4f243b13-1046-485b-b726-469b5a2b283f-reloader\") pod \"frr-k8s-7v4k6\" (UID: \"4f243b13-1046-485b-b726-469b5a2b283f\") " pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.744984 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4f243b13-1046-485b-b726-469b5a2b283f-frr-conf\") pod \"frr-k8s-7v4k6\" (UID: \"4f243b13-1046-485b-b726-469b5a2b283f\") " pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.745020 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4f243b13-1046-485b-b726-469b5a2b283f-frr-startup\") pod \"frr-k8s-7v4k6\" (UID: \"4f243b13-1046-485b-b726-469b5a2b283f\") " pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.745063 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl5fz\" (UniqueName: \"kubernetes.io/projected/4f243b13-1046-485b-b726-469b5a2b283f-kube-api-access-tl5fz\") pod \"frr-k8s-7v4k6\" (UID: \"4f243b13-1046-485b-b726-469b5a2b283f\") " pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.745087 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/69b8f173-b1f8-45e3-a6ae-ffb987fc709e-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-gk8sz\" (UID: \"69b8f173-b1f8-45e3-a6ae-ffb987fc709e\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-gk8sz" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.846923 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqj6q\" (UniqueName: \"kubernetes.io/projected/d8c2e94f-7eff-41dd-8e91-d6301ac9b84a-kube-api-access-cqj6q\") pod \"speaker-zfggl\" (UID: \"d8c2e94f-7eff-41dd-8e91-d6301ac9b84a\") " pod="metallb-system/speaker-zfggl" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.846994 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c-metrics-certs\") pod \"controller-7bb4cc7c98-lzh5g\" (UID: \"3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c\") " pod="metallb-system/controller-7bb4cc7c98-lzh5g" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.847022 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg7qh\" (UniqueName: \"kubernetes.io/projected/69b8f173-b1f8-45e3-a6ae-ffb987fc709e-kube-api-access-wg7qh\") pod \"frr-k8s-webhook-server-bcc4b6f68-gk8sz\" (UID: \"69b8f173-b1f8-45e3-a6ae-ffb987fc709e\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-gk8sz" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.847062 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4f243b13-1046-485b-b726-469b5a2b283f-metrics\") pod \"frr-k8s-7v4k6\" (UID: \"4f243b13-1046-485b-b726-469b5a2b283f\") " pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.847085 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4f243b13-1046-485b-b726-469b5a2b283f-metrics-certs\") pod \"frr-k8s-7v4k6\" (UID: \"4f243b13-1046-485b-b726-469b5a2b283f\") " pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.847117 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4f243b13-1046-485b-b726-469b5a2b283f-frr-sockets\") pod \"frr-k8s-7v4k6\" (UID: \"4f243b13-1046-485b-b726-469b5a2b283f\") " pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.847151 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c-cert\") pod \"controller-7bb4cc7c98-lzh5g\" (UID: \"3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c\") " pod="metallb-system/controller-7bb4cc7c98-lzh5g" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.847168 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4f243b13-1046-485b-b726-469b5a2b283f-reloader\") pod \"frr-k8s-7v4k6\" (UID: \"4f243b13-1046-485b-b726-469b5a2b283f\") " pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.847197 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d8c2e94f-7eff-41dd-8e91-d6301ac9b84a-memberlist\") pod \"speaker-zfggl\" (UID: \"d8c2e94f-7eff-41dd-8e91-d6301ac9b84a\") " pod="metallb-system/speaker-zfggl" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.847215 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d8c2e94f-7eff-41dd-8e91-d6301ac9b84a-metallb-excludel2\") pod \"speaker-zfggl\" (UID: \"d8c2e94f-7eff-41dd-8e91-d6301ac9b84a\") " pod="metallb-system/speaker-zfggl" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.847236 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4f243b13-1046-485b-b726-469b5a2b283f-frr-conf\") pod \"frr-k8s-7v4k6\" (UID: \"4f243b13-1046-485b-b726-469b5a2b283f\") " pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.847251 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbp6w\" (UniqueName: \"kubernetes.io/projected/3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c-kube-api-access-qbp6w\") pod \"controller-7bb4cc7c98-lzh5g\" (UID: \"3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c\") " pod="metallb-system/controller-7bb4cc7c98-lzh5g" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.847270 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d8c2e94f-7eff-41dd-8e91-d6301ac9b84a-metrics-certs\") pod \"speaker-zfggl\" (UID: \"d8c2e94f-7eff-41dd-8e91-d6301ac9b84a\") " pod="metallb-system/speaker-zfggl" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.847285 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4f243b13-1046-485b-b726-469b5a2b283f-frr-startup\") pod \"frr-k8s-7v4k6\" (UID: \"4f243b13-1046-485b-b726-469b5a2b283f\") " pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.847304 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl5fz\" (UniqueName: \"kubernetes.io/projected/4f243b13-1046-485b-b726-469b5a2b283f-kube-api-access-tl5fz\") pod \"frr-k8s-7v4k6\" (UID: \"4f243b13-1046-485b-b726-469b5a2b283f\") " pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.847320 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/69b8f173-b1f8-45e3-a6ae-ffb987fc709e-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-gk8sz\" (UID: \"69b8f173-b1f8-45e3-a6ae-ffb987fc709e\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-gk8sz" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.848627 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4f243b13-1046-485b-b726-469b5a2b283f-frr-conf\") pod \"frr-k8s-7v4k6\" (UID: \"4f243b13-1046-485b-b726-469b5a2b283f\") " pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.848721 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4f243b13-1046-485b-b726-469b5a2b283f-metrics\") pod \"frr-k8s-7v4k6\" (UID: \"4f243b13-1046-485b-b726-469b5a2b283f\") " pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.848902 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4f243b13-1046-485b-b726-469b5a2b283f-reloader\") pod \"frr-k8s-7v4k6\" (UID: \"4f243b13-1046-485b-b726-469b5a2b283f\") " pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.849183 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4f243b13-1046-485b-b726-469b5a2b283f-frr-sockets\") pod \"frr-k8s-7v4k6\" (UID: \"4f243b13-1046-485b-b726-469b5a2b283f\") " pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.849921 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4f243b13-1046-485b-b726-469b5a2b283f-frr-startup\") pod \"frr-k8s-7v4k6\" (UID: \"4f243b13-1046-485b-b726-469b5a2b283f\") " pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.856370 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4f243b13-1046-485b-b726-469b5a2b283f-metrics-certs\") pod \"frr-k8s-7v4k6\" (UID: \"4f243b13-1046-485b-b726-469b5a2b283f\") " pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.856618 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/69b8f173-b1f8-45e3-a6ae-ffb987fc709e-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-gk8sz\" (UID: \"69b8f173-b1f8-45e3-a6ae-ffb987fc709e\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-gk8sz" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.870805 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl5fz\" (UniqueName: \"kubernetes.io/projected/4f243b13-1046-485b-b726-469b5a2b283f-kube-api-access-tl5fz\") pod \"frr-k8s-7v4k6\" (UID: \"4f243b13-1046-485b-b726-469b5a2b283f\") " pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.878542 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg7qh\" (UniqueName: \"kubernetes.io/projected/69b8f173-b1f8-45e3-a6ae-ffb987fc709e-kube-api-access-wg7qh\") pod \"frr-k8s-webhook-server-bcc4b6f68-gk8sz\" (UID: \"69b8f173-b1f8-45e3-a6ae-ffb987fc709e\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-gk8sz" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.943731 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.948435 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c-metrics-certs\") pod \"controller-7bb4cc7c98-lzh5g\" (UID: \"3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c\") " pod="metallb-system/controller-7bb4cc7c98-lzh5g" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.949412 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c-cert\") pod \"controller-7bb4cc7c98-lzh5g\" (UID: \"3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c\") " pod="metallb-system/controller-7bb4cc7c98-lzh5g" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.949463 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d8c2e94f-7eff-41dd-8e91-d6301ac9b84a-memberlist\") pod \"speaker-zfggl\" (UID: \"d8c2e94f-7eff-41dd-8e91-d6301ac9b84a\") " pod="metallb-system/speaker-zfggl" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.949479 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d8c2e94f-7eff-41dd-8e91-d6301ac9b84a-metallb-excludel2\") pod \"speaker-zfggl\" (UID: \"d8c2e94f-7eff-41dd-8e91-d6301ac9b84a\") " pod="metallb-system/speaker-zfggl" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.949505 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbp6w\" (UniqueName: \"kubernetes.io/projected/3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c-kube-api-access-qbp6w\") pod \"controller-7bb4cc7c98-lzh5g\" (UID: \"3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c\") " pod="metallb-system/controller-7bb4cc7c98-lzh5g" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.949525 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d8c2e94f-7eff-41dd-8e91-d6301ac9b84a-metrics-certs\") pod \"speaker-zfggl\" (UID: \"d8c2e94f-7eff-41dd-8e91-d6301ac9b84a\") " pod="metallb-system/speaker-zfggl" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.949572 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqj6q\" (UniqueName: \"kubernetes.io/projected/d8c2e94f-7eff-41dd-8e91-d6301ac9b84a-kube-api-access-cqj6q\") pod \"speaker-zfggl\" (UID: \"d8c2e94f-7eff-41dd-8e91-d6301ac9b84a\") " pod="metallb-system/speaker-zfggl" Mar 21 08:42:35 crc kubenswrapper[4696]: E0321 08:42:35.949570 4696 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 21 08:42:35 crc kubenswrapper[4696]: E0321 08:42:35.949661 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d8c2e94f-7eff-41dd-8e91-d6301ac9b84a-memberlist podName:d8c2e94f-7eff-41dd-8e91-d6301ac9b84a nodeName:}" failed. No retries permitted until 2026-03-21 08:42:36.449646698 +0000 UTC m=+890.570527401 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d8c2e94f-7eff-41dd-8e91-d6301ac9b84a-memberlist") pod "speaker-zfggl" (UID: "d8c2e94f-7eff-41dd-8e91-d6301ac9b84a") : secret "metallb-memberlist" not found Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.952642 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-gk8sz" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.952771 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c-metrics-certs\") pod \"controller-7bb4cc7c98-lzh5g\" (UID: \"3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c\") " pod="metallb-system/controller-7bb4cc7c98-lzh5g" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.953168 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d8c2e94f-7eff-41dd-8e91-d6301ac9b84a-metrics-certs\") pod \"speaker-zfggl\" (UID: \"d8c2e94f-7eff-41dd-8e91-d6301ac9b84a\") " pod="metallb-system/speaker-zfggl" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.953269 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c-cert\") pod \"controller-7bb4cc7c98-lzh5g\" (UID: \"3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c\") " pod="metallb-system/controller-7bb4cc7c98-lzh5g" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.954304 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d8c2e94f-7eff-41dd-8e91-d6301ac9b84a-metallb-excludel2\") pod \"speaker-zfggl\" (UID: \"d8c2e94f-7eff-41dd-8e91-d6301ac9b84a\") " pod="metallb-system/speaker-zfggl" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.965382 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbp6w\" (UniqueName: \"kubernetes.io/projected/3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c-kube-api-access-qbp6w\") pod \"controller-7bb4cc7c98-lzh5g\" (UID: \"3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c\") " pod="metallb-system/controller-7bb4cc7c98-lzh5g" Mar 21 08:42:35 crc kubenswrapper[4696]: I0321 08:42:35.965525 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqj6q\" (UniqueName: \"kubernetes.io/projected/d8c2e94f-7eff-41dd-8e91-d6301ac9b84a-kube-api-access-cqj6q\") pod \"speaker-zfggl\" (UID: \"d8c2e94f-7eff-41dd-8e91-d6301ac9b84a\") " pod="metallb-system/speaker-zfggl" Mar 21 08:42:36 crc kubenswrapper[4696]: I0321 08:42:36.032171 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-lzh5g" Mar 21 08:42:36 crc kubenswrapper[4696]: I0321 08:42:36.405748 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-gk8sz"] Mar 21 08:42:36 crc kubenswrapper[4696]: W0321 08:42:36.414995 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69b8f173_b1f8_45e3_a6ae_ffb987fc709e.slice/crio-df6d292686c5b15e9fafae4486ca37d23111acfa7f0995761ac0d84f3f180c31 WatchSource:0}: Error finding container df6d292686c5b15e9fafae4486ca37d23111acfa7f0995761ac0d84f3f180c31: Status 404 returned error can't find the container with id df6d292686c5b15e9fafae4486ca37d23111acfa7f0995761ac0d84f3f180c31 Mar 21 08:42:36 crc kubenswrapper[4696]: I0321 08:42:36.464581 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-lzh5g"] Mar 21 08:42:36 crc kubenswrapper[4696]: W0321 08:42:36.472429 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a676a7c_ba6e_4b0c_bfdd_fbdfabdcf30c.slice/crio-7c9c8b9f6e39ae3c2058ab040122b4a455a1eb9b6e7be2650baf0214f294a33b WatchSource:0}: Error finding container 7c9c8b9f6e39ae3c2058ab040122b4a455a1eb9b6e7be2650baf0214f294a33b: Status 404 returned error can't find the container with id 7c9c8b9f6e39ae3c2058ab040122b4a455a1eb9b6e7be2650baf0214f294a33b Mar 21 08:42:36 crc kubenswrapper[4696]: I0321 08:42:36.478157 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d8c2e94f-7eff-41dd-8e91-d6301ac9b84a-memberlist\") pod \"speaker-zfggl\" (UID: \"d8c2e94f-7eff-41dd-8e91-d6301ac9b84a\") " pod="metallb-system/speaker-zfggl" Mar 21 08:42:36 crc kubenswrapper[4696]: E0321 08:42:36.478252 4696 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 21 08:42:36 crc kubenswrapper[4696]: E0321 08:42:36.478305 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d8c2e94f-7eff-41dd-8e91-d6301ac9b84a-memberlist podName:d8c2e94f-7eff-41dd-8e91-d6301ac9b84a nodeName:}" failed. No retries permitted until 2026-03-21 08:42:37.478290412 +0000 UTC m=+891.599171125 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d8c2e94f-7eff-41dd-8e91-d6301ac9b84a-memberlist") pod "speaker-zfggl" (UID: "d8c2e94f-7eff-41dd-8e91-d6301ac9b84a") : secret "metallb-memberlist" not found Mar 21 08:42:37 crc kubenswrapper[4696]: I0321 08:42:37.009446 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-lzh5g" event={"ID":"3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c","Type":"ContainerStarted","Data":"6177e36ab81fd37276491975155ba3cb5fe2e948be4182a64de2ba4854828f62"} Mar 21 08:42:37 crc kubenswrapper[4696]: I0321 08:42:37.009717 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-lzh5g" event={"ID":"3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c","Type":"ContainerStarted","Data":"cb8e5672a8a860dcf6311bdc5fa20ada1de21c1d8b9b92a9b21afab3a0b2e155"} Mar 21 08:42:37 crc kubenswrapper[4696]: I0321 08:42:37.009731 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-7bb4cc7c98-lzh5g" Mar 21 08:42:37 crc kubenswrapper[4696]: I0321 08:42:37.009741 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-lzh5g" event={"ID":"3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c","Type":"ContainerStarted","Data":"7c9c8b9f6e39ae3c2058ab040122b4a455a1eb9b6e7be2650baf0214f294a33b"} Mar 21 08:42:37 crc kubenswrapper[4696]: I0321 08:42:37.010467 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-gk8sz" event={"ID":"69b8f173-b1f8-45e3-a6ae-ffb987fc709e","Type":"ContainerStarted","Data":"df6d292686c5b15e9fafae4486ca37d23111acfa7f0995761ac0d84f3f180c31"} Mar 21 08:42:37 crc kubenswrapper[4696]: I0321 08:42:37.011711 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7v4k6" event={"ID":"4f243b13-1046-485b-b726-469b5a2b283f","Type":"ContainerStarted","Data":"b17c2360f32b7268e01ea0c0207ffa2cf4d17f9b907106898cc552d3016dacbe"} Mar 21 08:42:37 crc kubenswrapper[4696]: I0321 08:42:37.028410 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-7bb4cc7c98-lzh5g" podStartSLOduration=2.028366625 podStartE2EDuration="2.028366625s" podCreationTimestamp="2026-03-21 08:42:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:42:37.02408853 +0000 UTC m=+891.144969253" watchObservedRunningTime="2026-03-21 08:42:37.028366625 +0000 UTC m=+891.149247338" Mar 21 08:42:37 crc kubenswrapper[4696]: I0321 08:42:37.491347 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d8c2e94f-7eff-41dd-8e91-d6301ac9b84a-memberlist\") pod \"speaker-zfggl\" (UID: \"d8c2e94f-7eff-41dd-8e91-d6301ac9b84a\") " pod="metallb-system/speaker-zfggl" Mar 21 08:42:37 crc kubenswrapper[4696]: I0321 08:42:37.497016 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d8c2e94f-7eff-41dd-8e91-d6301ac9b84a-memberlist\") pod \"speaker-zfggl\" (UID: \"d8c2e94f-7eff-41dd-8e91-d6301ac9b84a\") " pod="metallb-system/speaker-zfggl" Mar 21 08:42:37 crc kubenswrapper[4696]: I0321 08:42:37.516453 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-zfggl" Mar 21 08:42:37 crc kubenswrapper[4696]: W0321 08:42:37.549186 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8c2e94f_7eff_41dd_8e91_d6301ac9b84a.slice/crio-888302ee653404928604ded7a95fefc639aa99131e3502c388cb1695bd0e6980 WatchSource:0}: Error finding container 888302ee653404928604ded7a95fefc639aa99131e3502c388cb1695bd0e6980: Status 404 returned error can't find the container with id 888302ee653404928604ded7a95fefc639aa99131e3502c388cb1695bd0e6980 Mar 21 08:42:38 crc kubenswrapper[4696]: I0321 08:42:38.037294 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zfggl" event={"ID":"d8c2e94f-7eff-41dd-8e91-d6301ac9b84a","Type":"ContainerStarted","Data":"a3150a9c56d042a5fb2e90893fd86b5c1bd995df4a335569ad02cc57c51c3ba2"} Mar 21 08:42:38 crc kubenswrapper[4696]: I0321 08:42:38.037606 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zfggl" event={"ID":"d8c2e94f-7eff-41dd-8e91-d6301ac9b84a","Type":"ContainerStarted","Data":"888302ee653404928604ded7a95fefc639aa99131e3502c388cb1695bd0e6980"} Mar 21 08:42:38 crc kubenswrapper[4696]: I0321 08:42:38.813711 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tvxpw"] Mar 21 08:42:38 crc kubenswrapper[4696]: I0321 08:42:38.815678 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tvxpw" Mar 21 08:42:38 crc kubenswrapper[4696]: I0321 08:42:38.829472 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tvxpw"] Mar 21 08:42:39 crc kubenswrapper[4696]: I0321 08:42:39.010140 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84d6411c-888b-4045-9a10-bed7ead4dfe5-catalog-content\") pod \"certified-operators-tvxpw\" (UID: \"84d6411c-888b-4045-9a10-bed7ead4dfe5\") " pod="openshift-marketplace/certified-operators-tvxpw" Mar 21 08:42:39 crc kubenswrapper[4696]: I0321 08:42:39.011594 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfpdj\" (UniqueName: \"kubernetes.io/projected/84d6411c-888b-4045-9a10-bed7ead4dfe5-kube-api-access-mfpdj\") pod \"certified-operators-tvxpw\" (UID: \"84d6411c-888b-4045-9a10-bed7ead4dfe5\") " pod="openshift-marketplace/certified-operators-tvxpw" Mar 21 08:42:39 crc kubenswrapper[4696]: I0321 08:42:39.011697 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84d6411c-888b-4045-9a10-bed7ead4dfe5-utilities\") pod \"certified-operators-tvxpw\" (UID: \"84d6411c-888b-4045-9a10-bed7ead4dfe5\") " pod="openshift-marketplace/certified-operators-tvxpw" Mar 21 08:42:39 crc kubenswrapper[4696]: I0321 08:42:39.052275 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zfggl" event={"ID":"d8c2e94f-7eff-41dd-8e91-d6301ac9b84a","Type":"ContainerStarted","Data":"5ec7770be4908dff827ffc5ad2db6d85a3e4c63fbe50a97f60431ea9b6ae314d"} Mar 21 08:42:39 crc kubenswrapper[4696]: I0321 08:42:39.052434 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-zfggl" Mar 21 08:42:39 crc kubenswrapper[4696]: I0321 08:42:39.076489 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-zfggl" podStartSLOduration=4.076468888 podStartE2EDuration="4.076468888s" podCreationTimestamp="2026-03-21 08:42:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:42:39.07137351 +0000 UTC m=+893.192254243" watchObservedRunningTime="2026-03-21 08:42:39.076468888 +0000 UTC m=+893.197349601" Mar 21 08:42:39 crc kubenswrapper[4696]: I0321 08:42:39.113169 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfpdj\" (UniqueName: \"kubernetes.io/projected/84d6411c-888b-4045-9a10-bed7ead4dfe5-kube-api-access-mfpdj\") pod \"certified-operators-tvxpw\" (UID: \"84d6411c-888b-4045-9a10-bed7ead4dfe5\") " pod="openshift-marketplace/certified-operators-tvxpw" Mar 21 08:42:39 crc kubenswrapper[4696]: I0321 08:42:39.113266 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84d6411c-888b-4045-9a10-bed7ead4dfe5-utilities\") pod \"certified-operators-tvxpw\" (UID: \"84d6411c-888b-4045-9a10-bed7ead4dfe5\") " pod="openshift-marketplace/certified-operators-tvxpw" Mar 21 08:42:39 crc kubenswrapper[4696]: I0321 08:42:39.113334 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84d6411c-888b-4045-9a10-bed7ead4dfe5-catalog-content\") pod \"certified-operators-tvxpw\" (UID: \"84d6411c-888b-4045-9a10-bed7ead4dfe5\") " pod="openshift-marketplace/certified-operators-tvxpw" Mar 21 08:42:39 crc kubenswrapper[4696]: I0321 08:42:39.114191 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84d6411c-888b-4045-9a10-bed7ead4dfe5-utilities\") pod \"certified-operators-tvxpw\" (UID: \"84d6411c-888b-4045-9a10-bed7ead4dfe5\") " pod="openshift-marketplace/certified-operators-tvxpw" Mar 21 08:42:39 crc kubenswrapper[4696]: I0321 08:42:39.114250 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84d6411c-888b-4045-9a10-bed7ead4dfe5-catalog-content\") pod \"certified-operators-tvxpw\" (UID: \"84d6411c-888b-4045-9a10-bed7ead4dfe5\") " pod="openshift-marketplace/certified-operators-tvxpw" Mar 21 08:42:39 crc kubenswrapper[4696]: I0321 08:42:39.145056 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfpdj\" (UniqueName: \"kubernetes.io/projected/84d6411c-888b-4045-9a10-bed7ead4dfe5-kube-api-access-mfpdj\") pod \"certified-operators-tvxpw\" (UID: \"84d6411c-888b-4045-9a10-bed7ead4dfe5\") " pod="openshift-marketplace/certified-operators-tvxpw" Mar 21 08:42:39 crc kubenswrapper[4696]: I0321 08:42:39.440370 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tvxpw" Mar 21 08:42:39 crc kubenswrapper[4696]: I0321 08:42:39.842361 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tvxpw"] Mar 21 08:42:40 crc kubenswrapper[4696]: I0321 08:42:40.066859 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tvxpw" event={"ID":"84d6411c-888b-4045-9a10-bed7ead4dfe5","Type":"ContainerStarted","Data":"4cfa69373f7aa608bda03542dcd29db1e3534bd298077f7642ec891d42e3f350"} Mar 21 08:42:40 crc kubenswrapper[4696]: I0321 08:42:40.066906 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tvxpw" event={"ID":"84d6411c-888b-4045-9a10-bed7ead4dfe5","Type":"ContainerStarted","Data":"9cf4254480a1a86499acf366419bac5a331021d2b28849237c0bdb98aa46a695"} Mar 21 08:42:41 crc kubenswrapper[4696]: I0321 08:42:41.074939 4696 generic.go:334] "Generic (PLEG): container finished" podID="84d6411c-888b-4045-9a10-bed7ead4dfe5" containerID="4cfa69373f7aa608bda03542dcd29db1e3534bd298077f7642ec891d42e3f350" exitCode=0 Mar 21 08:42:41 crc kubenswrapper[4696]: I0321 08:42:41.075279 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tvxpw" event={"ID":"84d6411c-888b-4045-9a10-bed7ead4dfe5","Type":"ContainerDied","Data":"4cfa69373f7aa608bda03542dcd29db1e3534bd298077f7642ec891d42e3f350"} Mar 21 08:42:45 crc kubenswrapper[4696]: I0321 08:42:45.101347 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-gk8sz" event={"ID":"69b8f173-b1f8-45e3-a6ae-ffb987fc709e","Type":"ContainerStarted","Data":"fc7749a49f61fdabf8c41e99566a8fb5de643020ceed79283f8de083f9ae6c38"} Mar 21 08:42:45 crc kubenswrapper[4696]: I0321 08:42:45.102759 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-gk8sz" Mar 21 08:42:45 crc kubenswrapper[4696]: I0321 08:42:45.105065 4696 generic.go:334] "Generic (PLEG): container finished" podID="4f243b13-1046-485b-b726-469b5a2b283f" containerID="059cb875535c8b1931c2ac0ba6131c0a05584f1d24dbd9bbe8bb216b6a89092e" exitCode=0 Mar 21 08:42:45 crc kubenswrapper[4696]: I0321 08:42:45.105121 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7v4k6" event={"ID":"4f243b13-1046-485b-b726-469b5a2b283f","Type":"ContainerDied","Data":"059cb875535c8b1931c2ac0ba6131c0a05584f1d24dbd9bbe8bb216b6a89092e"} Mar 21 08:42:45 crc kubenswrapper[4696]: I0321 08:42:45.119123 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-gk8sz" podStartSLOduration=1.831057767 podStartE2EDuration="10.11910614s" podCreationTimestamp="2026-03-21 08:42:35 +0000 UTC" firstStartedPulling="2026-03-21 08:42:36.417151601 +0000 UTC m=+890.538032314" lastFinishedPulling="2026-03-21 08:42:44.705199974 +0000 UTC m=+898.826080687" observedRunningTime="2026-03-21 08:42:45.117831456 +0000 UTC m=+899.238712169" watchObservedRunningTime="2026-03-21 08:42:45.11910614 +0000 UTC m=+899.239986853" Mar 21 08:42:45 crc kubenswrapper[4696]: I0321 08:42:45.147665 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qrvw9"] Mar 21 08:42:45 crc kubenswrapper[4696]: I0321 08:42:45.148982 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qrvw9" Mar 21 08:42:45 crc kubenswrapper[4696]: I0321 08:42:45.169594 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qrvw9"] Mar 21 08:42:45 crc kubenswrapper[4696]: I0321 08:42:45.318759 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsxnh\" (UniqueName: \"kubernetes.io/projected/55abc209-8a2b-4db3-8a0c-99688e2a454a-kube-api-access-xsxnh\") pod \"redhat-marketplace-qrvw9\" (UID: \"55abc209-8a2b-4db3-8a0c-99688e2a454a\") " pod="openshift-marketplace/redhat-marketplace-qrvw9" Mar 21 08:42:45 crc kubenswrapper[4696]: I0321 08:42:45.318877 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55abc209-8a2b-4db3-8a0c-99688e2a454a-utilities\") pod \"redhat-marketplace-qrvw9\" (UID: \"55abc209-8a2b-4db3-8a0c-99688e2a454a\") " pod="openshift-marketplace/redhat-marketplace-qrvw9" Mar 21 08:42:45 crc kubenswrapper[4696]: I0321 08:42:45.318920 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55abc209-8a2b-4db3-8a0c-99688e2a454a-catalog-content\") pod \"redhat-marketplace-qrvw9\" (UID: \"55abc209-8a2b-4db3-8a0c-99688e2a454a\") " pod="openshift-marketplace/redhat-marketplace-qrvw9" Mar 21 08:42:45 crc kubenswrapper[4696]: I0321 08:42:45.435839 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55abc209-8a2b-4db3-8a0c-99688e2a454a-catalog-content\") pod \"redhat-marketplace-qrvw9\" (UID: \"55abc209-8a2b-4db3-8a0c-99688e2a454a\") " pod="openshift-marketplace/redhat-marketplace-qrvw9" Mar 21 08:42:45 crc kubenswrapper[4696]: I0321 08:42:45.435934 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsxnh\" (UniqueName: \"kubernetes.io/projected/55abc209-8a2b-4db3-8a0c-99688e2a454a-kube-api-access-xsxnh\") pod \"redhat-marketplace-qrvw9\" (UID: \"55abc209-8a2b-4db3-8a0c-99688e2a454a\") " pod="openshift-marketplace/redhat-marketplace-qrvw9" Mar 21 08:42:45 crc kubenswrapper[4696]: I0321 08:42:45.435982 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55abc209-8a2b-4db3-8a0c-99688e2a454a-utilities\") pod \"redhat-marketplace-qrvw9\" (UID: \"55abc209-8a2b-4db3-8a0c-99688e2a454a\") " pod="openshift-marketplace/redhat-marketplace-qrvw9" Mar 21 08:42:45 crc kubenswrapper[4696]: I0321 08:42:45.436406 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55abc209-8a2b-4db3-8a0c-99688e2a454a-catalog-content\") pod \"redhat-marketplace-qrvw9\" (UID: \"55abc209-8a2b-4db3-8a0c-99688e2a454a\") " pod="openshift-marketplace/redhat-marketplace-qrvw9" Mar 21 08:42:45 crc kubenswrapper[4696]: I0321 08:42:45.436446 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55abc209-8a2b-4db3-8a0c-99688e2a454a-utilities\") pod \"redhat-marketplace-qrvw9\" (UID: \"55abc209-8a2b-4db3-8a0c-99688e2a454a\") " pod="openshift-marketplace/redhat-marketplace-qrvw9" Mar 21 08:42:45 crc kubenswrapper[4696]: I0321 08:42:45.465194 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsxnh\" (UniqueName: \"kubernetes.io/projected/55abc209-8a2b-4db3-8a0c-99688e2a454a-kube-api-access-xsxnh\") pod \"redhat-marketplace-qrvw9\" (UID: \"55abc209-8a2b-4db3-8a0c-99688e2a454a\") " pod="openshift-marketplace/redhat-marketplace-qrvw9" Mar 21 08:42:45 crc kubenswrapper[4696]: I0321 08:42:45.529356 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qrvw9" Mar 21 08:42:45 crc kubenswrapper[4696]: I0321 08:42:45.743068 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qrvw9"] Mar 21 08:42:46 crc kubenswrapper[4696]: I0321 08:42:46.037231 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-7bb4cc7c98-lzh5g" Mar 21 08:42:46 crc kubenswrapper[4696]: I0321 08:42:46.113468 4696 generic.go:334] "Generic (PLEG): container finished" podID="55abc209-8a2b-4db3-8a0c-99688e2a454a" containerID="64e157a8fad9fb3099872b3747e4c41b37ea0af2744eac50b52800e0bf8fd99b" exitCode=0 Mar 21 08:42:46 crc kubenswrapper[4696]: I0321 08:42:46.113558 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qrvw9" event={"ID":"55abc209-8a2b-4db3-8a0c-99688e2a454a","Type":"ContainerDied","Data":"64e157a8fad9fb3099872b3747e4c41b37ea0af2744eac50b52800e0bf8fd99b"} Mar 21 08:42:46 crc kubenswrapper[4696]: I0321 08:42:46.113587 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qrvw9" event={"ID":"55abc209-8a2b-4db3-8a0c-99688e2a454a","Type":"ContainerStarted","Data":"6f96b484a21c87b544601bf351d7b8572eaf11ab13627e600ba8da4c764fbe1b"} Mar 21 08:42:46 crc kubenswrapper[4696]: I0321 08:42:46.115707 4696 generic.go:334] "Generic (PLEG): container finished" podID="84d6411c-888b-4045-9a10-bed7ead4dfe5" containerID="c0ab54430ef698c4ff3402e3db6da204c280c6c39535df4c739bbbdca1ae032c" exitCode=0 Mar 21 08:42:46 crc kubenswrapper[4696]: I0321 08:42:46.115786 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tvxpw" event={"ID":"84d6411c-888b-4045-9a10-bed7ead4dfe5","Type":"ContainerDied","Data":"c0ab54430ef698c4ff3402e3db6da204c280c6c39535df4c739bbbdca1ae032c"} Mar 21 08:42:46 crc kubenswrapper[4696]: I0321 08:42:46.120864 4696 generic.go:334] "Generic (PLEG): container finished" podID="4f243b13-1046-485b-b726-469b5a2b283f" containerID="7cd989e4c3162bf072299932ecb28077f713f156aa903a3c4c74a1c3f1307792" exitCode=0 Mar 21 08:42:46 crc kubenswrapper[4696]: I0321 08:42:46.121072 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7v4k6" event={"ID":"4f243b13-1046-485b-b726-469b5a2b283f","Type":"ContainerDied","Data":"7cd989e4c3162bf072299932ecb28077f713f156aa903a3c4c74a1c3f1307792"} Mar 21 08:42:47 crc kubenswrapper[4696]: I0321 08:42:47.128760 4696 generic.go:334] "Generic (PLEG): container finished" podID="4f243b13-1046-485b-b726-469b5a2b283f" containerID="48343643ebd6a117fc97d18938b58e375bd52dc1df5f90195f7258c076bbb2a3" exitCode=0 Mar 21 08:42:47 crc kubenswrapper[4696]: I0321 08:42:47.128864 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7v4k6" event={"ID":"4f243b13-1046-485b-b726-469b5a2b283f","Type":"ContainerDied","Data":"48343643ebd6a117fc97d18938b58e375bd52dc1df5f90195f7258c076bbb2a3"} Mar 21 08:42:47 crc kubenswrapper[4696]: I0321 08:42:47.156364 4696 scope.go:117] "RemoveContainer" containerID="b2fb2bf44f23b85fb9dce261b53e63b215dcba4f7829a2354c3221ed6d402cc3" Mar 21 08:42:47 crc kubenswrapper[4696]: I0321 08:42:47.519867 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-zfggl" Mar 21 08:42:48 crc kubenswrapper[4696]: I0321 08:42:48.137256 4696 generic.go:334] "Generic (PLEG): container finished" podID="55abc209-8a2b-4db3-8a0c-99688e2a454a" containerID="29c8c39a28f33b564635e5f9ca6cf7f480ca70bef8a2f95a069d67ec2695dc3d" exitCode=0 Mar 21 08:42:48 crc kubenswrapper[4696]: I0321 08:42:48.137336 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qrvw9" event={"ID":"55abc209-8a2b-4db3-8a0c-99688e2a454a","Type":"ContainerDied","Data":"29c8c39a28f33b564635e5f9ca6cf7f480ca70bef8a2f95a069d67ec2695dc3d"} Mar 21 08:42:48 crc kubenswrapper[4696]: I0321 08:42:48.142108 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tvxpw" event={"ID":"84d6411c-888b-4045-9a10-bed7ead4dfe5","Type":"ContainerStarted","Data":"6a4a350df0c0304502c1d05189036a1fc362b53af12b1a1cca8faddf6663016e"} Mar 21 08:42:48 crc kubenswrapper[4696]: I0321 08:42:48.148055 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7v4k6" event={"ID":"4f243b13-1046-485b-b726-469b5a2b283f","Type":"ContainerStarted","Data":"ff9edd1c744948353bf5ca209563e8ca52cf1c5d4e661197fe79fd6a2b875536"} Mar 21 08:42:48 crc kubenswrapper[4696]: I0321 08:42:48.148095 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7v4k6" event={"ID":"4f243b13-1046-485b-b726-469b5a2b283f","Type":"ContainerStarted","Data":"d66b8b18ca67a0a1678d24ccffc2ba7644c6813f759707f8ed68d0f996451247"} Mar 21 08:42:48 crc kubenswrapper[4696]: I0321 08:42:48.148110 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7v4k6" event={"ID":"4f243b13-1046-485b-b726-469b5a2b283f","Type":"ContainerStarted","Data":"76cb9534a1233095ec5db6c898c4dcf176581afaa1df5faf73c31739ec266d06"} Mar 21 08:42:48 crc kubenswrapper[4696]: I0321 08:42:48.193697 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tvxpw" podStartSLOduration=7.229159431 podStartE2EDuration="10.193676428s" podCreationTimestamp="2026-03-21 08:42:38 +0000 UTC" firstStartedPulling="2026-03-21 08:42:44.282534681 +0000 UTC m=+898.403415394" lastFinishedPulling="2026-03-21 08:42:47.247051678 +0000 UTC m=+901.367932391" observedRunningTime="2026-03-21 08:42:48.183505154 +0000 UTC m=+902.304385897" watchObservedRunningTime="2026-03-21 08:42:48.193676428 +0000 UTC m=+902.314557141" Mar 21 08:42:49 crc kubenswrapper[4696]: I0321 08:42:49.158504 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7v4k6" event={"ID":"4f243b13-1046-485b-b726-469b5a2b283f","Type":"ContainerStarted","Data":"b15cd44eb273fbdf9e06c5cb80954e721142d64942f9ed0f39bec7cab4bb81da"} Mar 21 08:42:49 crc kubenswrapper[4696]: I0321 08:42:49.158876 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:49 crc kubenswrapper[4696]: I0321 08:42:49.158889 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7v4k6" event={"ID":"4f243b13-1046-485b-b726-469b5a2b283f","Type":"ContainerStarted","Data":"4a1ce70b3064efc7bb82425ac56fadcf55e2ee4f0656779e106f959e77e6df25"} Mar 21 08:42:49 crc kubenswrapper[4696]: I0321 08:42:49.158898 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7v4k6" event={"ID":"4f243b13-1046-485b-b726-469b5a2b283f","Type":"ContainerStarted","Data":"3f9bfc06f41bbc095945e430f68915f3f39d34186bc62f1f9ffbc74b487246ee"} Mar 21 08:42:49 crc kubenswrapper[4696]: I0321 08:42:49.160719 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qrvw9" event={"ID":"55abc209-8a2b-4db3-8a0c-99688e2a454a","Type":"ContainerStarted","Data":"f7ec7364ad3c60f1a8911ec575dcbae5f2cd24439fcd68c6af45b78383e4edf8"} Mar 21 08:42:49 crc kubenswrapper[4696]: I0321 08:42:49.182730 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-7v4k6" podStartSLOduration=5.708002724 podStartE2EDuration="14.182711115s" podCreationTimestamp="2026-03-21 08:42:35 +0000 UTC" firstStartedPulling="2026-03-21 08:42:36.207320596 +0000 UTC m=+890.328201309" lastFinishedPulling="2026-03-21 08:42:44.682028987 +0000 UTC m=+898.802909700" observedRunningTime="2026-03-21 08:42:49.180472124 +0000 UTC m=+903.301352837" watchObservedRunningTime="2026-03-21 08:42:49.182711115 +0000 UTC m=+903.303591828" Mar 21 08:42:49 crc kubenswrapper[4696]: I0321 08:42:49.441241 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tvxpw" Mar 21 08:42:49 crc kubenswrapper[4696]: I0321 08:42:49.441309 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tvxpw" Mar 21 08:42:50 crc kubenswrapper[4696]: I0321 08:42:50.486520 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-tvxpw" podUID="84d6411c-888b-4045-9a10-bed7ead4dfe5" containerName="registry-server" probeResult="failure" output=< Mar 21 08:42:50 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Mar 21 08:42:50 crc kubenswrapper[4696]: > Mar 21 08:42:50 crc kubenswrapper[4696]: I0321 08:42:50.944188 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:50 crc kubenswrapper[4696]: I0321 08:42:50.985306 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:42:51 crc kubenswrapper[4696]: I0321 08:42:51.014871 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qrvw9" podStartSLOduration=3.588266753 podStartE2EDuration="6.014855845s" podCreationTimestamp="2026-03-21 08:42:45 +0000 UTC" firstStartedPulling="2026-03-21 08:42:46.115139215 +0000 UTC m=+900.236019928" lastFinishedPulling="2026-03-21 08:42:48.541728307 +0000 UTC m=+902.662609020" observedRunningTime="2026-03-21 08:42:49.199305463 +0000 UTC m=+903.320186196" watchObservedRunningTime="2026-03-21 08:42:51.014855845 +0000 UTC m=+905.135736558" Mar 21 08:42:54 crc kubenswrapper[4696]: I0321 08:42:54.719233 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-k24x8"] Mar 21 08:42:54 crc kubenswrapper[4696]: I0321 08:42:54.720263 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-k24x8" Mar 21 08:42:54 crc kubenswrapper[4696]: I0321 08:42:54.722282 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 21 08:42:54 crc kubenswrapper[4696]: I0321 08:42:54.722512 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 21 08:42:54 crc kubenswrapper[4696]: I0321 08:42:54.723389 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-9m5pb" Mar 21 08:42:54 crc kubenswrapper[4696]: I0321 08:42:54.735658 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-k24x8"] Mar 21 08:42:54 crc kubenswrapper[4696]: I0321 08:42:54.753708 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs76x\" (UniqueName: \"kubernetes.io/projected/2e3b5f4d-52bd-4d2c-b629-74fc14699179-kube-api-access-fs76x\") pod \"openstack-operator-index-k24x8\" (UID: \"2e3b5f4d-52bd-4d2c-b629-74fc14699179\") " pod="openstack-operators/openstack-operator-index-k24x8" Mar 21 08:42:54 crc kubenswrapper[4696]: I0321 08:42:54.854670 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs76x\" (UniqueName: \"kubernetes.io/projected/2e3b5f4d-52bd-4d2c-b629-74fc14699179-kube-api-access-fs76x\") pod \"openstack-operator-index-k24x8\" (UID: \"2e3b5f4d-52bd-4d2c-b629-74fc14699179\") " pod="openstack-operators/openstack-operator-index-k24x8" Mar 21 08:42:54 crc kubenswrapper[4696]: I0321 08:42:54.872734 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs76x\" (UniqueName: \"kubernetes.io/projected/2e3b5f4d-52bd-4d2c-b629-74fc14699179-kube-api-access-fs76x\") pod \"openstack-operator-index-k24x8\" (UID: \"2e3b5f4d-52bd-4d2c-b629-74fc14699179\") " pod="openstack-operators/openstack-operator-index-k24x8" Mar 21 08:42:55 crc kubenswrapper[4696]: I0321 08:42:55.036608 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-k24x8" Mar 21 08:42:55 crc kubenswrapper[4696]: I0321 08:42:55.457389 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-k24x8"] Mar 21 08:42:55 crc kubenswrapper[4696]: I0321 08:42:55.530550 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qrvw9" Mar 21 08:42:55 crc kubenswrapper[4696]: I0321 08:42:55.530628 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qrvw9" Mar 21 08:42:55 crc kubenswrapper[4696]: I0321 08:42:55.576105 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qrvw9" Mar 21 08:42:55 crc kubenswrapper[4696]: I0321 08:42:55.957679 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-gk8sz" Mar 21 08:42:56 crc kubenswrapper[4696]: I0321 08:42:56.208638 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-k24x8" event={"ID":"2e3b5f4d-52bd-4d2c-b629-74fc14699179","Type":"ContainerStarted","Data":"1415c9412683ae1b224dfca11777d0ea54f039011183a5e097aa663f6c82c0d9"} Mar 21 08:42:56 crc kubenswrapper[4696]: I0321 08:42:56.251883 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qrvw9" Mar 21 08:42:59 crc kubenswrapper[4696]: I0321 08:42:59.481373 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tvxpw" Mar 21 08:42:59 crc kubenswrapper[4696]: I0321 08:42:59.525713 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qrvw9"] Mar 21 08:42:59 crc kubenswrapper[4696]: I0321 08:42:59.526199 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qrvw9" podUID="55abc209-8a2b-4db3-8a0c-99688e2a454a" containerName="registry-server" containerID="cri-o://f7ec7364ad3c60f1a8911ec575dcbae5f2cd24439fcd68c6af45b78383e4edf8" gracePeriod=2 Mar 21 08:42:59 crc kubenswrapper[4696]: I0321 08:42:59.556864 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tvxpw" Mar 21 08:43:01 crc kubenswrapper[4696]: I0321 08:43:01.240247 4696 generic.go:334] "Generic (PLEG): container finished" podID="55abc209-8a2b-4db3-8a0c-99688e2a454a" containerID="f7ec7364ad3c60f1a8911ec575dcbae5f2cd24439fcd68c6af45b78383e4edf8" exitCode=0 Mar 21 08:43:01 crc kubenswrapper[4696]: I0321 08:43:01.240324 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qrvw9" event={"ID":"55abc209-8a2b-4db3-8a0c-99688e2a454a","Type":"ContainerDied","Data":"f7ec7364ad3c60f1a8911ec575dcbae5f2cd24439fcd68c6af45b78383e4edf8"} Mar 21 08:43:01 crc kubenswrapper[4696]: I0321 08:43:01.856942 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qrvw9" Mar 21 08:43:01 crc kubenswrapper[4696]: I0321 08:43:01.958921 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55abc209-8a2b-4db3-8a0c-99688e2a454a-catalog-content\") pod \"55abc209-8a2b-4db3-8a0c-99688e2a454a\" (UID: \"55abc209-8a2b-4db3-8a0c-99688e2a454a\") " Mar 21 08:43:01 crc kubenswrapper[4696]: I0321 08:43:01.959018 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55abc209-8a2b-4db3-8a0c-99688e2a454a-utilities\") pod \"55abc209-8a2b-4db3-8a0c-99688e2a454a\" (UID: \"55abc209-8a2b-4db3-8a0c-99688e2a454a\") " Mar 21 08:43:01 crc kubenswrapper[4696]: I0321 08:43:01.959085 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsxnh\" (UniqueName: \"kubernetes.io/projected/55abc209-8a2b-4db3-8a0c-99688e2a454a-kube-api-access-xsxnh\") pod \"55abc209-8a2b-4db3-8a0c-99688e2a454a\" (UID: \"55abc209-8a2b-4db3-8a0c-99688e2a454a\") " Mar 21 08:43:01 crc kubenswrapper[4696]: I0321 08:43:01.959999 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55abc209-8a2b-4db3-8a0c-99688e2a454a-utilities" (OuterVolumeSpecName: "utilities") pod "55abc209-8a2b-4db3-8a0c-99688e2a454a" (UID: "55abc209-8a2b-4db3-8a0c-99688e2a454a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:43:01 crc kubenswrapper[4696]: I0321 08:43:01.964400 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55abc209-8a2b-4db3-8a0c-99688e2a454a-kube-api-access-xsxnh" (OuterVolumeSpecName: "kube-api-access-xsxnh") pod "55abc209-8a2b-4db3-8a0c-99688e2a454a" (UID: "55abc209-8a2b-4db3-8a0c-99688e2a454a"). InnerVolumeSpecName "kube-api-access-xsxnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:43:01 crc kubenswrapper[4696]: I0321 08:43:01.982202 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55abc209-8a2b-4db3-8a0c-99688e2a454a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55abc209-8a2b-4db3-8a0c-99688e2a454a" (UID: "55abc209-8a2b-4db3-8a0c-99688e2a454a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:43:02 crc kubenswrapper[4696]: I0321 08:43:02.060114 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsxnh\" (UniqueName: \"kubernetes.io/projected/55abc209-8a2b-4db3-8a0c-99688e2a454a-kube-api-access-xsxnh\") on node \"crc\" DevicePath \"\"" Mar 21 08:43:02 crc kubenswrapper[4696]: I0321 08:43:02.060150 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55abc209-8a2b-4db3-8a0c-99688e2a454a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 08:43:02 crc kubenswrapper[4696]: I0321 08:43:02.060165 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55abc209-8a2b-4db3-8a0c-99688e2a454a-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 08:43:02 crc kubenswrapper[4696]: I0321 08:43:02.251215 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qrvw9" event={"ID":"55abc209-8a2b-4db3-8a0c-99688e2a454a","Type":"ContainerDied","Data":"6f96b484a21c87b544601bf351d7b8572eaf11ab13627e600ba8da4c764fbe1b"} Mar 21 08:43:02 crc kubenswrapper[4696]: I0321 08:43:02.251286 4696 scope.go:117] "RemoveContainer" containerID="f7ec7364ad3c60f1a8911ec575dcbae5f2cd24439fcd68c6af45b78383e4edf8" Mar 21 08:43:02 crc kubenswrapper[4696]: I0321 08:43:02.251482 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qrvw9" Mar 21 08:43:02 crc kubenswrapper[4696]: I0321 08:43:02.282966 4696 scope.go:117] "RemoveContainer" containerID="29c8c39a28f33b564635e5f9ca6cf7f480ca70bef8a2f95a069d67ec2695dc3d" Mar 21 08:43:02 crc kubenswrapper[4696]: I0321 08:43:02.291105 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qrvw9"] Mar 21 08:43:02 crc kubenswrapper[4696]: I0321 08:43:02.295424 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qrvw9"] Mar 21 08:43:02 crc kubenswrapper[4696]: I0321 08:43:02.320055 4696 scope.go:117] "RemoveContainer" containerID="64e157a8fad9fb3099872b3747e4c41b37ea0af2744eac50b52800e0bf8fd99b" Mar 21 08:43:02 crc kubenswrapper[4696]: I0321 08:43:02.547699 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55abc209-8a2b-4db3-8a0c-99688e2a454a" path="/var/lib/kubelet/pods/55abc209-8a2b-4db3-8a0c-99688e2a454a/volumes" Mar 21 08:43:02 crc kubenswrapper[4696]: I0321 08:43:02.917509 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tvxpw"] Mar 21 08:43:02 crc kubenswrapper[4696]: I0321 08:43:02.918084 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tvxpw" podUID="84d6411c-888b-4045-9a10-bed7ead4dfe5" containerName="registry-server" containerID="cri-o://6a4a350df0c0304502c1d05189036a1fc362b53af12b1a1cca8faddf6663016e" gracePeriod=2 Mar 21 08:43:03 crc kubenswrapper[4696]: I0321 08:43:03.261807 4696 generic.go:334] "Generic (PLEG): container finished" podID="84d6411c-888b-4045-9a10-bed7ead4dfe5" containerID="6a4a350df0c0304502c1d05189036a1fc362b53af12b1a1cca8faddf6663016e" exitCode=0 Mar 21 08:43:03 crc kubenswrapper[4696]: I0321 08:43:03.261860 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tvxpw" event={"ID":"84d6411c-888b-4045-9a10-bed7ead4dfe5","Type":"ContainerDied","Data":"6a4a350df0c0304502c1d05189036a1fc362b53af12b1a1cca8faddf6663016e"} Mar 21 08:43:03 crc kubenswrapper[4696]: I0321 08:43:03.511685 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tvxpw" Mar 21 08:43:03 crc kubenswrapper[4696]: I0321 08:43:03.681132 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84d6411c-888b-4045-9a10-bed7ead4dfe5-utilities\") pod \"84d6411c-888b-4045-9a10-bed7ead4dfe5\" (UID: \"84d6411c-888b-4045-9a10-bed7ead4dfe5\") " Mar 21 08:43:03 crc kubenswrapper[4696]: I0321 08:43:03.681216 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84d6411c-888b-4045-9a10-bed7ead4dfe5-catalog-content\") pod \"84d6411c-888b-4045-9a10-bed7ead4dfe5\" (UID: \"84d6411c-888b-4045-9a10-bed7ead4dfe5\") " Mar 21 08:43:03 crc kubenswrapper[4696]: I0321 08:43:03.681266 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfpdj\" (UniqueName: \"kubernetes.io/projected/84d6411c-888b-4045-9a10-bed7ead4dfe5-kube-api-access-mfpdj\") pod \"84d6411c-888b-4045-9a10-bed7ead4dfe5\" (UID: \"84d6411c-888b-4045-9a10-bed7ead4dfe5\") " Mar 21 08:43:03 crc kubenswrapper[4696]: I0321 08:43:03.682515 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84d6411c-888b-4045-9a10-bed7ead4dfe5-utilities" (OuterVolumeSpecName: "utilities") pod "84d6411c-888b-4045-9a10-bed7ead4dfe5" (UID: "84d6411c-888b-4045-9a10-bed7ead4dfe5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:43:03 crc kubenswrapper[4696]: I0321 08:43:03.687157 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84d6411c-888b-4045-9a10-bed7ead4dfe5-kube-api-access-mfpdj" (OuterVolumeSpecName: "kube-api-access-mfpdj") pod "84d6411c-888b-4045-9a10-bed7ead4dfe5" (UID: "84d6411c-888b-4045-9a10-bed7ead4dfe5"). InnerVolumeSpecName "kube-api-access-mfpdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:43:03 crc kubenswrapper[4696]: I0321 08:43:03.733984 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84d6411c-888b-4045-9a10-bed7ead4dfe5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84d6411c-888b-4045-9a10-bed7ead4dfe5" (UID: "84d6411c-888b-4045-9a10-bed7ead4dfe5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:43:03 crc kubenswrapper[4696]: I0321 08:43:03.782201 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfpdj\" (UniqueName: \"kubernetes.io/projected/84d6411c-888b-4045-9a10-bed7ead4dfe5-kube-api-access-mfpdj\") on node \"crc\" DevicePath \"\"" Mar 21 08:43:03 crc kubenswrapper[4696]: I0321 08:43:03.782235 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84d6411c-888b-4045-9a10-bed7ead4dfe5-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 08:43:03 crc kubenswrapper[4696]: I0321 08:43:03.782245 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84d6411c-888b-4045-9a10-bed7ead4dfe5-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 08:43:04 crc kubenswrapper[4696]: I0321 08:43:04.284190 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tvxpw" event={"ID":"84d6411c-888b-4045-9a10-bed7ead4dfe5","Type":"ContainerDied","Data":"9cf4254480a1a86499acf366419bac5a331021d2b28849237c0bdb98aa46a695"} Mar 21 08:43:04 crc kubenswrapper[4696]: I0321 08:43:04.284258 4696 scope.go:117] "RemoveContainer" containerID="6a4a350df0c0304502c1d05189036a1fc362b53af12b1a1cca8faddf6663016e" Mar 21 08:43:04 crc kubenswrapper[4696]: I0321 08:43:04.284266 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tvxpw" Mar 21 08:43:04 crc kubenswrapper[4696]: I0321 08:43:04.314617 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tvxpw"] Mar 21 08:43:04 crc kubenswrapper[4696]: I0321 08:43:04.319721 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tvxpw"] Mar 21 08:43:04 crc kubenswrapper[4696]: I0321 08:43:04.547643 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84d6411c-888b-4045-9a10-bed7ead4dfe5" path="/var/lib/kubelet/pods/84d6411c-888b-4045-9a10-bed7ead4dfe5/volumes" Mar 21 08:43:04 crc kubenswrapper[4696]: I0321 08:43:04.949123 4696 scope.go:117] "RemoveContainer" containerID="c0ab54430ef698c4ff3402e3db6da204c280c6c39535df4c739bbbdca1ae032c" Mar 21 08:43:05 crc kubenswrapper[4696]: I0321 08:43:05.661799 4696 scope.go:117] "RemoveContainer" containerID="4cfa69373f7aa608bda03542dcd29db1e3534bd298077f7642ec891d42e3f350" Mar 21 08:43:05 crc kubenswrapper[4696]: I0321 08:43:05.949789 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-7v4k6" Mar 21 08:43:09 crc kubenswrapper[4696]: I0321 08:43:09.317565 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-k24x8" event={"ID":"2e3b5f4d-52bd-4d2c-b629-74fc14699179","Type":"ContainerStarted","Data":"41526c41cf2fbc4ad6ee9d1ad8825333d9c40291367b6c09811b881392ec3c52"} Mar 21 08:43:09 crc kubenswrapper[4696]: I0321 08:43:09.335139 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-k24x8" podStartSLOduration=2.334474455 podStartE2EDuration="15.335120386s" podCreationTimestamp="2026-03-21 08:42:54 +0000 UTC" firstStartedPulling="2026-03-21 08:42:55.468546853 +0000 UTC m=+909.589427566" lastFinishedPulling="2026-03-21 08:43:08.469192784 +0000 UTC m=+922.590073497" observedRunningTime="2026-03-21 08:43:09.331810037 +0000 UTC m=+923.452690760" watchObservedRunningTime="2026-03-21 08:43:09.335120386 +0000 UTC m=+923.456001099" Mar 21 08:43:15 crc kubenswrapper[4696]: I0321 08:43:15.037315 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-k24x8" Mar 21 08:43:15 crc kubenswrapper[4696]: I0321 08:43:15.037665 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-k24x8" Mar 21 08:43:15 crc kubenswrapper[4696]: I0321 08:43:15.067655 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-k24x8" Mar 21 08:43:15 crc kubenswrapper[4696]: I0321 08:43:15.379911 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-k24x8" Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.509191 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t"] Mar 21 08:43:23 crc kubenswrapper[4696]: E0321 08:43:23.509913 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d6411c-888b-4045-9a10-bed7ead4dfe5" containerName="registry-server" Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.509925 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d6411c-888b-4045-9a10-bed7ead4dfe5" containerName="registry-server" Mar 21 08:43:23 crc kubenswrapper[4696]: E0321 08:43:23.509938 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55abc209-8a2b-4db3-8a0c-99688e2a454a" containerName="extract-utilities" Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.509945 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="55abc209-8a2b-4db3-8a0c-99688e2a454a" containerName="extract-utilities" Mar 21 08:43:23 crc kubenswrapper[4696]: E0321 08:43:23.509952 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55abc209-8a2b-4db3-8a0c-99688e2a454a" containerName="registry-server" Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.509958 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="55abc209-8a2b-4db3-8a0c-99688e2a454a" containerName="registry-server" Mar 21 08:43:23 crc kubenswrapper[4696]: E0321 08:43:23.509970 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d6411c-888b-4045-9a10-bed7ead4dfe5" containerName="extract-utilities" Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.509975 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d6411c-888b-4045-9a10-bed7ead4dfe5" containerName="extract-utilities" Mar 21 08:43:23 crc kubenswrapper[4696]: E0321 08:43:23.509984 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d6411c-888b-4045-9a10-bed7ead4dfe5" containerName="extract-content" Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.509989 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d6411c-888b-4045-9a10-bed7ead4dfe5" containerName="extract-content" Mar 21 08:43:23 crc kubenswrapper[4696]: E0321 08:43:23.510000 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55abc209-8a2b-4db3-8a0c-99688e2a454a" containerName="extract-content" Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.510006 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="55abc209-8a2b-4db3-8a0c-99688e2a454a" containerName="extract-content" Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.510105 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d6411c-888b-4045-9a10-bed7ead4dfe5" containerName="registry-server" Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.510113 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="55abc209-8a2b-4db3-8a0c-99688e2a454a" containerName="registry-server" Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.510977 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t" Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.512729 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-4bhkb" Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.522714 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t"] Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.648288 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/01e5ce30-73e2-4e41-9d25-6ac3c67f9361-util\") pod \"364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t\" (UID: \"01e5ce30-73e2-4e41-9d25-6ac3c67f9361\") " pod="openstack-operators/364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t" Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.648344 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwggj\" (UniqueName: \"kubernetes.io/projected/01e5ce30-73e2-4e41-9d25-6ac3c67f9361-kube-api-access-nwggj\") pod \"364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t\" (UID: \"01e5ce30-73e2-4e41-9d25-6ac3c67f9361\") " pod="openstack-operators/364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t" Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.648397 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/01e5ce30-73e2-4e41-9d25-6ac3c67f9361-bundle\") pod \"364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t\" (UID: \"01e5ce30-73e2-4e41-9d25-6ac3c67f9361\") " pod="openstack-operators/364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t" Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.749498 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/01e5ce30-73e2-4e41-9d25-6ac3c67f9361-util\") pod \"364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t\" (UID: \"01e5ce30-73e2-4e41-9d25-6ac3c67f9361\") " pod="openstack-operators/364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t" Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.749575 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwggj\" (UniqueName: \"kubernetes.io/projected/01e5ce30-73e2-4e41-9d25-6ac3c67f9361-kube-api-access-nwggj\") pod \"364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t\" (UID: \"01e5ce30-73e2-4e41-9d25-6ac3c67f9361\") " pod="openstack-operators/364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t" Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.749612 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/01e5ce30-73e2-4e41-9d25-6ac3c67f9361-bundle\") pod \"364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t\" (UID: \"01e5ce30-73e2-4e41-9d25-6ac3c67f9361\") " pod="openstack-operators/364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t" Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.750065 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/01e5ce30-73e2-4e41-9d25-6ac3c67f9361-util\") pod \"364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t\" (UID: \"01e5ce30-73e2-4e41-9d25-6ac3c67f9361\") " pod="openstack-operators/364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t" Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.750081 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/01e5ce30-73e2-4e41-9d25-6ac3c67f9361-bundle\") pod \"364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t\" (UID: \"01e5ce30-73e2-4e41-9d25-6ac3c67f9361\") " pod="openstack-operators/364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t" Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.768294 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwggj\" (UniqueName: \"kubernetes.io/projected/01e5ce30-73e2-4e41-9d25-6ac3c67f9361-kube-api-access-nwggj\") pod \"364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t\" (UID: \"01e5ce30-73e2-4e41-9d25-6ac3c67f9361\") " pod="openstack-operators/364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t" Mar 21 08:43:23 crc kubenswrapper[4696]: I0321 08:43:23.830369 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t" Mar 21 08:43:24 crc kubenswrapper[4696]: I0321 08:43:24.230575 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t"] Mar 21 08:43:24 crc kubenswrapper[4696]: W0321 08:43:24.232980 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01e5ce30_73e2_4e41_9d25_6ac3c67f9361.slice/crio-35450976599f000ebf4774bef60bcda56f4ac46914f37dcb53555ba00c8194fb WatchSource:0}: Error finding container 35450976599f000ebf4774bef60bcda56f4ac46914f37dcb53555ba00c8194fb: Status 404 returned error can't find the container with id 35450976599f000ebf4774bef60bcda56f4ac46914f37dcb53555ba00c8194fb Mar 21 08:43:24 crc kubenswrapper[4696]: I0321 08:43:24.414836 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t" event={"ID":"01e5ce30-73e2-4e41-9d25-6ac3c67f9361","Type":"ContainerStarted","Data":"35450976599f000ebf4774bef60bcda56f4ac46914f37dcb53555ba00c8194fb"} Mar 21 08:43:25 crc kubenswrapper[4696]: I0321 08:43:25.421586 4696 generic.go:334] "Generic (PLEG): container finished" podID="01e5ce30-73e2-4e41-9d25-6ac3c67f9361" containerID="86c7add5bf2940402093a9a8762ca7a711ad343dfaf1e7c69a13d3b14c9b21d8" exitCode=0 Mar 21 08:43:25 crc kubenswrapper[4696]: I0321 08:43:25.421861 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t" event={"ID":"01e5ce30-73e2-4e41-9d25-6ac3c67f9361","Type":"ContainerDied","Data":"86c7add5bf2940402093a9a8762ca7a711ad343dfaf1e7c69a13d3b14c9b21d8"} Mar 21 08:43:26 crc kubenswrapper[4696]: I0321 08:43:26.429669 4696 generic.go:334] "Generic (PLEG): container finished" podID="01e5ce30-73e2-4e41-9d25-6ac3c67f9361" containerID="16e90da077f803dd32915a0048efc507d12f2b43c398207cc058fc31db860ae0" exitCode=0 Mar 21 08:43:26 crc kubenswrapper[4696]: I0321 08:43:26.429697 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t" event={"ID":"01e5ce30-73e2-4e41-9d25-6ac3c67f9361","Type":"ContainerDied","Data":"16e90da077f803dd32915a0048efc507d12f2b43c398207cc058fc31db860ae0"} Mar 21 08:43:27 crc kubenswrapper[4696]: I0321 08:43:27.454454 4696 generic.go:334] "Generic (PLEG): container finished" podID="01e5ce30-73e2-4e41-9d25-6ac3c67f9361" containerID="acd1efbe11a917b6973b03488cd1be971a5114687845c48860e4d38de86084de" exitCode=0 Mar 21 08:43:27 crc kubenswrapper[4696]: I0321 08:43:27.454528 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t" event={"ID":"01e5ce30-73e2-4e41-9d25-6ac3c67f9361","Type":"ContainerDied","Data":"acd1efbe11a917b6973b03488cd1be971a5114687845c48860e4d38de86084de"} Mar 21 08:43:28 crc kubenswrapper[4696]: I0321 08:43:28.695261 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t" Mar 21 08:43:28 crc kubenswrapper[4696]: I0321 08:43:28.820307 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/01e5ce30-73e2-4e41-9d25-6ac3c67f9361-bundle\") pod \"01e5ce30-73e2-4e41-9d25-6ac3c67f9361\" (UID: \"01e5ce30-73e2-4e41-9d25-6ac3c67f9361\") " Mar 21 08:43:28 crc kubenswrapper[4696]: I0321 08:43:28.820371 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/01e5ce30-73e2-4e41-9d25-6ac3c67f9361-util\") pod \"01e5ce30-73e2-4e41-9d25-6ac3c67f9361\" (UID: \"01e5ce30-73e2-4e41-9d25-6ac3c67f9361\") " Mar 21 08:43:28 crc kubenswrapper[4696]: I0321 08:43:28.820398 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwggj\" (UniqueName: \"kubernetes.io/projected/01e5ce30-73e2-4e41-9d25-6ac3c67f9361-kube-api-access-nwggj\") pod \"01e5ce30-73e2-4e41-9d25-6ac3c67f9361\" (UID: \"01e5ce30-73e2-4e41-9d25-6ac3c67f9361\") " Mar 21 08:43:28 crc kubenswrapper[4696]: I0321 08:43:28.820966 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01e5ce30-73e2-4e41-9d25-6ac3c67f9361-bundle" (OuterVolumeSpecName: "bundle") pod "01e5ce30-73e2-4e41-9d25-6ac3c67f9361" (UID: "01e5ce30-73e2-4e41-9d25-6ac3c67f9361"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:43:28 crc kubenswrapper[4696]: I0321 08:43:28.824855 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01e5ce30-73e2-4e41-9d25-6ac3c67f9361-kube-api-access-nwggj" (OuterVolumeSpecName: "kube-api-access-nwggj") pod "01e5ce30-73e2-4e41-9d25-6ac3c67f9361" (UID: "01e5ce30-73e2-4e41-9d25-6ac3c67f9361"). InnerVolumeSpecName "kube-api-access-nwggj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:43:28 crc kubenswrapper[4696]: I0321 08:43:28.833982 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01e5ce30-73e2-4e41-9d25-6ac3c67f9361-util" (OuterVolumeSpecName: "util") pod "01e5ce30-73e2-4e41-9d25-6ac3c67f9361" (UID: "01e5ce30-73e2-4e41-9d25-6ac3c67f9361"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:43:28 crc kubenswrapper[4696]: I0321 08:43:28.922023 4696 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/01e5ce30-73e2-4e41-9d25-6ac3c67f9361-util\") on node \"crc\" DevicePath \"\"" Mar 21 08:43:28 crc kubenswrapper[4696]: I0321 08:43:28.922060 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwggj\" (UniqueName: \"kubernetes.io/projected/01e5ce30-73e2-4e41-9d25-6ac3c67f9361-kube-api-access-nwggj\") on node \"crc\" DevicePath \"\"" Mar 21 08:43:28 crc kubenswrapper[4696]: I0321 08:43:28.922070 4696 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/01e5ce30-73e2-4e41-9d25-6ac3c67f9361-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:43:29 crc kubenswrapper[4696]: I0321 08:43:29.469771 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t" event={"ID":"01e5ce30-73e2-4e41-9d25-6ac3c67f9361","Type":"ContainerDied","Data":"35450976599f000ebf4774bef60bcda56f4ac46914f37dcb53555ba00c8194fb"} Mar 21 08:43:29 crc kubenswrapper[4696]: I0321 08:43:29.469812 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35450976599f000ebf4774bef60bcda56f4ac46914f37dcb53555ba00c8194fb" Mar 21 08:43:29 crc kubenswrapper[4696]: I0321 08:43:29.469884 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t" Mar 21 08:43:30 crc kubenswrapper[4696]: I0321 08:43:30.341486 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:43:30 crc kubenswrapper[4696]: I0321 08:43:30.341550 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:43:36 crc kubenswrapper[4696]: I0321 08:43:36.432599 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-888f66487-rsc92"] Mar 21 08:43:36 crc kubenswrapper[4696]: E0321 08:43:36.433473 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01e5ce30-73e2-4e41-9d25-6ac3c67f9361" containerName="util" Mar 21 08:43:36 crc kubenswrapper[4696]: I0321 08:43:36.433487 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="01e5ce30-73e2-4e41-9d25-6ac3c67f9361" containerName="util" Mar 21 08:43:36 crc kubenswrapper[4696]: E0321 08:43:36.433511 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01e5ce30-73e2-4e41-9d25-6ac3c67f9361" containerName="extract" Mar 21 08:43:36 crc kubenswrapper[4696]: I0321 08:43:36.433518 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="01e5ce30-73e2-4e41-9d25-6ac3c67f9361" containerName="extract" Mar 21 08:43:36 crc kubenswrapper[4696]: E0321 08:43:36.433530 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01e5ce30-73e2-4e41-9d25-6ac3c67f9361" containerName="pull" Mar 21 08:43:36 crc kubenswrapper[4696]: I0321 08:43:36.433536 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="01e5ce30-73e2-4e41-9d25-6ac3c67f9361" containerName="pull" Mar 21 08:43:36 crc kubenswrapper[4696]: I0321 08:43:36.433665 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="01e5ce30-73e2-4e41-9d25-6ac3c67f9361" containerName="extract" Mar 21 08:43:36 crc kubenswrapper[4696]: I0321 08:43:36.434222 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-888f66487-rsc92" Mar 21 08:43:36 crc kubenswrapper[4696]: I0321 08:43:36.436064 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-9nx65" Mar 21 08:43:36 crc kubenswrapper[4696]: I0321 08:43:36.584774 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-888f66487-rsc92"] Mar 21 08:43:36 crc kubenswrapper[4696]: I0321 08:43:36.615155 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km8zl\" (UniqueName: \"kubernetes.io/projected/bfd4f311-01db-41ce-8357-93860d30b442-kube-api-access-km8zl\") pod \"openstack-operator-controller-init-888f66487-rsc92\" (UID: \"bfd4f311-01db-41ce-8357-93860d30b442\") " pod="openstack-operators/openstack-operator-controller-init-888f66487-rsc92" Mar 21 08:43:36 crc kubenswrapper[4696]: I0321 08:43:36.716234 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km8zl\" (UniqueName: \"kubernetes.io/projected/bfd4f311-01db-41ce-8357-93860d30b442-kube-api-access-km8zl\") pod \"openstack-operator-controller-init-888f66487-rsc92\" (UID: \"bfd4f311-01db-41ce-8357-93860d30b442\") " pod="openstack-operators/openstack-operator-controller-init-888f66487-rsc92" Mar 21 08:43:36 crc kubenswrapper[4696]: I0321 08:43:36.735667 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km8zl\" (UniqueName: \"kubernetes.io/projected/bfd4f311-01db-41ce-8357-93860d30b442-kube-api-access-km8zl\") pod \"openstack-operator-controller-init-888f66487-rsc92\" (UID: \"bfd4f311-01db-41ce-8357-93860d30b442\") " pod="openstack-operators/openstack-operator-controller-init-888f66487-rsc92" Mar 21 08:43:36 crc kubenswrapper[4696]: I0321 08:43:36.752197 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-888f66487-rsc92" Mar 21 08:43:37 crc kubenswrapper[4696]: I0321 08:43:37.187389 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-888f66487-rsc92"] Mar 21 08:43:37 crc kubenswrapper[4696]: I0321 08:43:37.515506 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-888f66487-rsc92" event={"ID":"bfd4f311-01db-41ce-8357-93860d30b442","Type":"ContainerStarted","Data":"7401748ec7e94cbd6c908f28887a24a58989106ccee46c0dd81a968b2fa27c3b"} Mar 21 08:43:42 crc kubenswrapper[4696]: I0321 08:43:42.545661 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-888f66487-rsc92" event={"ID":"bfd4f311-01db-41ce-8357-93860d30b442","Type":"ContainerStarted","Data":"fe5a025c456898566d681b39438c03965c19451542e52f51ee95d8af64b396ff"} Mar 21 08:43:42 crc kubenswrapper[4696]: I0321 08:43:42.546275 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-888f66487-rsc92" Mar 21 08:43:42 crc kubenswrapper[4696]: I0321 08:43:42.571021 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-888f66487-rsc92" podStartSLOduration=1.922004713 podStartE2EDuration="6.570998594s" podCreationTimestamp="2026-03-21 08:43:36 +0000 UTC" firstStartedPulling="2026-03-21 08:43:37.195112047 +0000 UTC m=+951.315992760" lastFinishedPulling="2026-03-21 08:43:41.844105928 +0000 UTC m=+955.964986641" observedRunningTime="2026-03-21 08:43:42.56491452 +0000 UTC m=+956.685795233" watchObservedRunningTime="2026-03-21 08:43:42.570998594 +0000 UTC m=+956.691879307" Mar 21 08:43:46 crc kubenswrapper[4696]: I0321 08:43:46.756227 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-888f66487-rsc92" Mar 21 08:44:00 crc kubenswrapper[4696]: I0321 08:44:00.160832 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568044-g82fg"] Mar 21 08:44:00 crc kubenswrapper[4696]: I0321 08:44:00.162210 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568044-g82fg" Mar 21 08:44:00 crc kubenswrapper[4696]: I0321 08:44:00.164402 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 08:44:00 crc kubenswrapper[4696]: I0321 08:44:00.165052 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 08:44:00 crc kubenswrapper[4696]: I0321 08:44:00.165083 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 08:44:00 crc kubenswrapper[4696]: I0321 08:44:00.188921 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568044-g82fg"] Mar 21 08:44:00 crc kubenswrapper[4696]: I0321 08:44:00.232626 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jwcl\" (UniqueName: \"kubernetes.io/projected/703e51f2-bf70-45c9-ac92-3d96a4d49d18-kube-api-access-8jwcl\") pod \"auto-csr-approver-29568044-g82fg\" (UID: \"703e51f2-bf70-45c9-ac92-3d96a4d49d18\") " pod="openshift-infra/auto-csr-approver-29568044-g82fg" Mar 21 08:44:00 crc kubenswrapper[4696]: I0321 08:44:00.333797 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jwcl\" (UniqueName: \"kubernetes.io/projected/703e51f2-bf70-45c9-ac92-3d96a4d49d18-kube-api-access-8jwcl\") pod \"auto-csr-approver-29568044-g82fg\" (UID: \"703e51f2-bf70-45c9-ac92-3d96a4d49d18\") " pod="openshift-infra/auto-csr-approver-29568044-g82fg" Mar 21 08:44:00 crc kubenswrapper[4696]: I0321 08:44:00.341637 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:44:00 crc kubenswrapper[4696]: I0321 08:44:00.341981 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:44:00 crc kubenswrapper[4696]: I0321 08:44:00.370250 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jwcl\" (UniqueName: \"kubernetes.io/projected/703e51f2-bf70-45c9-ac92-3d96a4d49d18-kube-api-access-8jwcl\") pod \"auto-csr-approver-29568044-g82fg\" (UID: \"703e51f2-bf70-45c9-ac92-3d96a4d49d18\") " pod="openshift-infra/auto-csr-approver-29568044-g82fg" Mar 21 08:44:00 crc kubenswrapper[4696]: I0321 08:44:00.480222 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568044-g82fg" Mar 21 08:44:00 crc kubenswrapper[4696]: I0321 08:44:00.888367 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568044-g82fg"] Mar 21 08:44:01 crc kubenswrapper[4696]: I0321 08:44:01.658119 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568044-g82fg" event={"ID":"703e51f2-bf70-45c9-ac92-3d96a4d49d18","Type":"ContainerStarted","Data":"c9c87193f87835c91aa23b671ae6cb8427ee6658daf0bfea6c018eb589a7c9f2"} Mar 21 08:44:02 crc kubenswrapper[4696]: I0321 08:44:02.665557 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568044-g82fg" event={"ID":"703e51f2-bf70-45c9-ac92-3d96a4d49d18","Type":"ContainerStarted","Data":"25b39e1721f34b8f1fc2ea8ab54b44f3e42a606c54d75eec4d1ef997119951b3"} Mar 21 08:44:02 crc kubenswrapper[4696]: I0321 08:44:02.680180 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29568044-g82fg" podStartSLOduration=1.152808374 podStartE2EDuration="2.680164831s" podCreationTimestamp="2026-03-21 08:44:00 +0000 UTC" firstStartedPulling="2026-03-21 08:44:00.892684249 +0000 UTC m=+975.013564962" lastFinishedPulling="2026-03-21 08:44:02.420040706 +0000 UTC m=+976.540921419" observedRunningTime="2026-03-21 08:44:02.679405809 +0000 UTC m=+976.800286532" watchObservedRunningTime="2026-03-21 08:44:02.680164831 +0000 UTC m=+976.801045544" Mar 21 08:44:03 crc kubenswrapper[4696]: I0321 08:44:03.673724 4696 generic.go:334] "Generic (PLEG): container finished" podID="703e51f2-bf70-45c9-ac92-3d96a4d49d18" containerID="25b39e1721f34b8f1fc2ea8ab54b44f3e42a606c54d75eec4d1ef997119951b3" exitCode=0 Mar 21 08:44:03 crc kubenswrapper[4696]: I0321 08:44:03.673772 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568044-g82fg" event={"ID":"703e51f2-bf70-45c9-ac92-3d96a4d49d18","Type":"ContainerDied","Data":"25b39e1721f34b8f1fc2ea8ab54b44f3e42a606c54d75eec4d1ef997119951b3"} Mar 21 08:44:05 crc kubenswrapper[4696]: I0321 08:44:05.015865 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568044-g82fg" Mar 21 08:44:05 crc kubenswrapper[4696]: I0321 08:44:05.127943 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jwcl\" (UniqueName: \"kubernetes.io/projected/703e51f2-bf70-45c9-ac92-3d96a4d49d18-kube-api-access-8jwcl\") pod \"703e51f2-bf70-45c9-ac92-3d96a4d49d18\" (UID: \"703e51f2-bf70-45c9-ac92-3d96a4d49d18\") " Mar 21 08:44:05 crc kubenswrapper[4696]: I0321 08:44:05.132736 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/703e51f2-bf70-45c9-ac92-3d96a4d49d18-kube-api-access-8jwcl" (OuterVolumeSpecName: "kube-api-access-8jwcl") pod "703e51f2-bf70-45c9-ac92-3d96a4d49d18" (UID: "703e51f2-bf70-45c9-ac92-3d96a4d49d18"). InnerVolumeSpecName "kube-api-access-8jwcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:44:05 crc kubenswrapper[4696]: I0321 08:44:05.229331 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jwcl\" (UniqueName: \"kubernetes.io/projected/703e51f2-bf70-45c9-ac92-3d96a4d49d18-kube-api-access-8jwcl\") on node \"crc\" DevicePath \"\"" Mar 21 08:44:05 crc kubenswrapper[4696]: I0321 08:44:05.686941 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568044-g82fg" event={"ID":"703e51f2-bf70-45c9-ac92-3d96a4d49d18","Type":"ContainerDied","Data":"c9c87193f87835c91aa23b671ae6cb8427ee6658daf0bfea6c018eb589a7c9f2"} Mar 21 08:44:05 crc kubenswrapper[4696]: I0321 08:44:05.686984 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9c87193f87835c91aa23b671ae6cb8427ee6658daf0bfea6c018eb589a7c9f2" Mar 21 08:44:05 crc kubenswrapper[4696]: I0321 08:44:05.686989 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568044-g82fg" Mar 21 08:44:05 crc kubenswrapper[4696]: I0321 08:44:05.752620 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568038-gss7n"] Mar 21 08:44:05 crc kubenswrapper[4696]: I0321 08:44:05.756834 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568038-gss7n"] Mar 21 08:44:06 crc kubenswrapper[4696]: I0321 08:44:06.541962 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a11bb77-b5ab-41fb-aea9-5d7de836af84" path="/var/lib/kubelet/pods/9a11bb77-b5ab-41fb-aea9-5d7de836af84/volumes" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.473874 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5cfd84c587-ldj9m"] Mar 21 08:44:11 crc kubenswrapper[4696]: E0321 08:44:11.474712 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="703e51f2-bf70-45c9-ac92-3d96a4d49d18" containerName="oc" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.474729 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="703e51f2-bf70-45c9-ac92-3d96a4d49d18" containerName="oc" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.474877 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="703e51f2-bf70-45c9-ac92-3d96a4d49d18" containerName="oc" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.475407 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5cfd84c587-ldj9m" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.477247 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-r25r6" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.487768 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6d77645966-qsd27"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.488626 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6d77645966-qsd27" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.492847 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-ks475" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.507123 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6d77645966-qsd27"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.512768 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qtcd\" (UniqueName: \"kubernetes.io/projected/53a9e616-007d-4c0c-8b0a-799d7188d9ab-kube-api-access-6qtcd\") pod \"barbican-operator-controller-manager-5cfd84c587-ldj9m\" (UID: \"53a9e616-007d-4c0c-8b0a-799d7188d9ab\") " pod="openstack-operators/barbican-operator-controller-manager-5cfd84c587-ldj9m" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.558998 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-6cc65c69fc-m8ffd"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.568661 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6cc65c69fc-m8ffd" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.575381 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-m6xwz" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.576892 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7d559dcdbd-zxgmc"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.587001 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7d559dcdbd-zxgmc" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.591316 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-86gl4" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.600850 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-66dd9d474d-qf5kd"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.601597 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-66dd9d474d-qf5kd" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.607786 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-6dqr8" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.617577 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwbmn\" (UniqueName: \"kubernetes.io/projected/9ce080a1-65e5-4235-b181-448e8987c590-kube-api-access-gwbmn\") pod \"glance-operator-controller-manager-7d559dcdbd-zxgmc\" (UID: \"9ce080a1-65e5-4235-b181-448e8987c590\") " pod="openstack-operators/glance-operator-controller-manager-7d559dcdbd-zxgmc" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.617665 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qtcd\" (UniqueName: \"kubernetes.io/projected/53a9e616-007d-4c0c-8b0a-799d7188d9ab-kube-api-access-6qtcd\") pod \"barbican-operator-controller-manager-5cfd84c587-ldj9m\" (UID: \"53a9e616-007d-4c0c-8b0a-799d7188d9ab\") " pod="openstack-operators/barbican-operator-controller-manager-5cfd84c587-ldj9m" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.617689 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sbtg\" (UniqueName: \"kubernetes.io/projected/3cd62d0b-d740-4bef-ac3c-993b5ee05d40-kube-api-access-6sbtg\") pod \"cinder-operator-controller-manager-6d77645966-qsd27\" (UID: \"3cd62d0b-d740-4bef-ac3c-993b5ee05d40\") " pod="openstack-operators/cinder-operator-controller-manager-6d77645966-qsd27" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.617719 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntf5f\" (UniqueName: \"kubernetes.io/projected/530eca0b-da29-4883-ad3a-1faa1563c65d-kube-api-access-ntf5f\") pod \"designate-operator-controller-manager-6cc65c69fc-m8ffd\" (UID: \"530eca0b-da29-4883-ad3a-1faa1563c65d\") " pod="openstack-operators/designate-operator-controller-manager-6cc65c69fc-m8ffd" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.636173 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6cc65c69fc-m8ffd"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.641160 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7d559dcdbd-zxgmc"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.642120 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qtcd\" (UniqueName: \"kubernetes.io/projected/53a9e616-007d-4c0c-8b0a-799d7188d9ab-kube-api-access-6qtcd\") pod \"barbican-operator-controller-manager-5cfd84c587-ldj9m\" (UID: \"53a9e616-007d-4c0c-8b0a-799d7188d9ab\") " pod="openstack-operators/barbican-operator-controller-manager-5cfd84c587-ldj9m" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.649674 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-64dc66d669-8685h"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.650502 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-64dc66d669-8685h" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.652863 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-5595c7d6ff-7cg24"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.653701 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5595c7d6ff-7cg24" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.657991 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5cfd84c587-ldj9m"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.668274 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6b77b7676d-vhnjl"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.669296 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6b77b7676d-vhnjl" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.673854 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5595c7d6ff-7cg24"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.675084 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-rmxtq" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.680273 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-2rxbp" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.680522 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.680530 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-r9dj9" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.681415 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-66dd9d474d-qf5kd"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.689013 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-76b87776c9-rcz2g"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.694490 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-76b87776c9-rcz2g" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.694634 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6b77b7676d-vhnjl"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.696201 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-zj7z9" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.718721 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz2cp\" (UniqueName: \"kubernetes.io/projected/58ec1c2d-43a6-4582-9a21-f1083777ffe5-kube-api-access-dz2cp\") pod \"heat-operator-controller-manager-66dd9d474d-qf5kd\" (UID: \"58ec1c2d-43a6-4582-9a21-f1083777ffe5\") " pod="openstack-operators/heat-operator-controller-manager-66dd9d474d-qf5kd" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.718877 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwbmn\" (UniqueName: \"kubernetes.io/projected/9ce080a1-65e5-4235-b181-448e8987c590-kube-api-access-gwbmn\") pod \"glance-operator-controller-manager-7d559dcdbd-zxgmc\" (UID: \"9ce080a1-65e5-4235-b181-448e8987c590\") " pod="openstack-operators/glance-operator-controller-manager-7d559dcdbd-zxgmc" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.718920 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0048e137-52d0-476b-845a-4c42829de1e2-cert\") pod \"infra-operator-controller-manager-5595c7d6ff-7cg24\" (UID: \"0048e137-52d0-476b-845a-4c42829de1e2\") " pod="openstack-operators/infra-operator-controller-manager-5595c7d6ff-7cg24" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.718944 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g689j\" (UniqueName: \"kubernetes.io/projected/9b05c8a9-af90-4497-94f3-8fed890ab3b0-kube-api-access-g689j\") pod \"horizon-operator-controller-manager-64dc66d669-8685h\" (UID: \"9b05c8a9-af90-4497-94f3-8fed890ab3b0\") " pod="openstack-operators/horizon-operator-controller-manager-64dc66d669-8685h" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.718971 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9tp5\" (UniqueName: \"kubernetes.io/projected/74fc6c6a-6eb4-4459-aed9-62073e0472e4-kube-api-access-g9tp5\") pod \"ironic-operator-controller-manager-6b77b7676d-vhnjl\" (UID: \"74fc6c6a-6eb4-4459-aed9-62073e0472e4\") " pod="openstack-operators/ironic-operator-controller-manager-6b77b7676d-vhnjl" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.718992 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sbtg\" (UniqueName: \"kubernetes.io/projected/3cd62d0b-d740-4bef-ac3c-993b5ee05d40-kube-api-access-6sbtg\") pod \"cinder-operator-controller-manager-6d77645966-qsd27\" (UID: \"3cd62d0b-d740-4bef-ac3c-993b5ee05d40\") " pod="openstack-operators/cinder-operator-controller-manager-6d77645966-qsd27" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.719010 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9npw\" (UniqueName: \"kubernetes.io/projected/0048e137-52d0-476b-845a-4c42829de1e2-kube-api-access-d9npw\") pod \"infra-operator-controller-manager-5595c7d6ff-7cg24\" (UID: \"0048e137-52d0-476b-845a-4c42829de1e2\") " pod="openstack-operators/infra-operator-controller-manager-5595c7d6ff-7cg24" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.719032 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntf5f\" (UniqueName: \"kubernetes.io/projected/530eca0b-da29-4883-ad3a-1faa1563c65d-kube-api-access-ntf5f\") pod \"designate-operator-controller-manager-6cc65c69fc-m8ffd\" (UID: \"530eca0b-da29-4883-ad3a-1faa1563c65d\") " pod="openstack-operators/designate-operator-controller-manager-6cc65c69fc-m8ffd" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.720110 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-64dc66d669-8685h"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.727394 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-76b87776c9-rcz2g"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.734007 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-fbf7bbb96-745cq"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.741262 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-fbf7bbb96-745cq" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.747415 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-wjbdd" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.762391 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-fbf7bbb96-745cq"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.770465 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntf5f\" (UniqueName: \"kubernetes.io/projected/530eca0b-da29-4883-ad3a-1faa1563c65d-kube-api-access-ntf5f\") pod \"designate-operator-controller-manager-6cc65c69fc-m8ffd\" (UID: \"530eca0b-da29-4883-ad3a-1faa1563c65d\") " pod="openstack-operators/designate-operator-controller-manager-6cc65c69fc-m8ffd" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.770534 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6f5b7bcd4-pzgkp"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.771328 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6f5b7bcd4-pzgkp" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.776607 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sbtg\" (UniqueName: \"kubernetes.io/projected/3cd62d0b-d740-4bef-ac3c-993b5ee05d40-kube-api-access-6sbtg\") pod \"cinder-operator-controller-manager-6d77645966-qsd27\" (UID: \"3cd62d0b-d740-4bef-ac3c-993b5ee05d40\") " pod="openstack-operators/cinder-operator-controller-manager-6d77645966-qsd27" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.779686 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-s6gdp" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.779865 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6f5b7bcd4-pzgkp"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.789385 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwbmn\" (UniqueName: \"kubernetes.io/projected/9ce080a1-65e5-4235-b181-448e8987c590-kube-api-access-gwbmn\") pod \"glance-operator-controller-manager-7d559dcdbd-zxgmc\" (UID: \"9ce080a1-65e5-4235-b181-448e8987c590\") " pod="openstack-operators/glance-operator-controller-manager-7d559dcdbd-zxgmc" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.790984 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6744dd545c-mvqjq"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.791758 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6744dd545c-mvqjq" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.796222 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5cfd84c587-ldj9m" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.798439 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-7jrfm" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.810190 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6d77645966-qsd27" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.822644 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6744dd545c-mvqjq"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.823270 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69czt\" (UniqueName: \"kubernetes.io/projected/f9ca1637-123c-47eb-8183-5394a0ba6f89-kube-api-access-69czt\") pod \"neutron-operator-controller-manager-6744dd545c-mvqjq\" (UID: \"f9ca1637-123c-47eb-8183-5394a0ba6f89\") " pod="openstack-operators/neutron-operator-controller-manager-6744dd545c-mvqjq" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.823324 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz2cp\" (UniqueName: \"kubernetes.io/projected/58ec1c2d-43a6-4582-9a21-f1083777ffe5-kube-api-access-dz2cp\") pod \"heat-operator-controller-manager-66dd9d474d-qf5kd\" (UID: \"58ec1c2d-43a6-4582-9a21-f1083777ffe5\") " pod="openstack-operators/heat-operator-controller-manager-66dd9d474d-qf5kd" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.823343 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lndx9\" (UniqueName: \"kubernetes.io/projected/8e08ed5d-23b6-4066-a86f-3019f8cb34c8-kube-api-access-lndx9\") pod \"manila-operator-controller-manager-fbf7bbb96-745cq\" (UID: \"8e08ed5d-23b6-4066-a86f-3019f8cb34c8\") " pod="openstack-operators/manila-operator-controller-manager-fbf7bbb96-745cq" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.823368 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvlt8\" (UniqueName: \"kubernetes.io/projected/a4bbdaf9-e480-4cd1-9bf8-304cf5266949-kube-api-access-bvlt8\") pod \"mariadb-operator-controller-manager-6f5b7bcd4-pzgkp\" (UID: \"a4bbdaf9-e480-4cd1-9bf8-304cf5266949\") " pod="openstack-operators/mariadb-operator-controller-manager-6f5b7bcd4-pzgkp" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.823408 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0048e137-52d0-476b-845a-4c42829de1e2-cert\") pod \"infra-operator-controller-manager-5595c7d6ff-7cg24\" (UID: \"0048e137-52d0-476b-845a-4c42829de1e2\") " pod="openstack-operators/infra-operator-controller-manager-5595c7d6ff-7cg24" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.823430 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g689j\" (UniqueName: \"kubernetes.io/projected/9b05c8a9-af90-4497-94f3-8fed890ab3b0-kube-api-access-g689j\") pod \"horizon-operator-controller-manager-64dc66d669-8685h\" (UID: \"9b05c8a9-af90-4497-94f3-8fed890ab3b0\") " pod="openstack-operators/horizon-operator-controller-manager-64dc66d669-8685h" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.823450 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9tp5\" (UniqueName: \"kubernetes.io/projected/74fc6c6a-6eb4-4459-aed9-62073e0472e4-kube-api-access-g9tp5\") pod \"ironic-operator-controller-manager-6b77b7676d-vhnjl\" (UID: \"74fc6c6a-6eb4-4459-aed9-62073e0472e4\") " pod="openstack-operators/ironic-operator-controller-manager-6b77b7676d-vhnjl" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.823471 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvtr5\" (UniqueName: \"kubernetes.io/projected/c31077f0-de38-4de7-8421-a5ea7ac3ab84-kube-api-access-gvtr5\") pod \"keystone-operator-controller-manager-76b87776c9-rcz2g\" (UID: \"c31077f0-de38-4de7-8421-a5ea7ac3ab84\") " pod="openstack-operators/keystone-operator-controller-manager-76b87776c9-rcz2g" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.823487 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9npw\" (UniqueName: \"kubernetes.io/projected/0048e137-52d0-476b-845a-4c42829de1e2-kube-api-access-d9npw\") pod \"infra-operator-controller-manager-5595c7d6ff-7cg24\" (UID: \"0048e137-52d0-476b-845a-4c42829de1e2\") " pod="openstack-operators/infra-operator-controller-manager-5595c7d6ff-7cg24" Mar 21 08:44:11 crc kubenswrapper[4696]: E0321 08:44:11.823971 4696 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 21 08:44:11 crc kubenswrapper[4696]: E0321 08:44:11.824012 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0048e137-52d0-476b-845a-4c42829de1e2-cert podName:0048e137-52d0-476b-845a-4c42829de1e2 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:12.323995989 +0000 UTC m=+986.444876692 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0048e137-52d0-476b-845a-4c42829de1e2-cert") pod "infra-operator-controller-manager-5595c7d6ff-7cg24" (UID: "0048e137-52d0-476b-845a-4c42829de1e2") : secret "infra-operator-webhook-server-cert" not found Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.824689 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-bc5c78db9-c9skl"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.825488 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-bc5c78db9-c9skl" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.828616 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-56f74467c6-fzv5q"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.829397 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-56f74467c6-fzv5q" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.837044 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-56f74467c6-fzv5q"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.842307 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-tkftj" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.842585 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-hjm6x" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.843760 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-bc5c78db9-c9skl"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.863365 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g689j\" (UniqueName: \"kubernetes.io/projected/9b05c8a9-af90-4497-94f3-8fed890ab3b0-kube-api-access-g689j\") pod \"horizon-operator-controller-manager-64dc66d669-8685h\" (UID: \"9b05c8a9-af90-4497-94f3-8fed890ab3b0\") " pod="openstack-operators/horizon-operator-controller-manager-64dc66d669-8685h" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.864914 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9tp5\" (UniqueName: \"kubernetes.io/projected/74fc6c6a-6eb4-4459-aed9-62073e0472e4-kube-api-access-g9tp5\") pod \"ironic-operator-controller-manager-6b77b7676d-vhnjl\" (UID: \"74fc6c6a-6eb4-4459-aed9-62073e0472e4\") " pod="openstack-operators/ironic-operator-controller-manager-6b77b7676d-vhnjl" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.878575 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9npw\" (UniqueName: \"kubernetes.io/projected/0048e137-52d0-476b-845a-4c42829de1e2-kube-api-access-d9npw\") pod \"infra-operator-controller-manager-5595c7d6ff-7cg24\" (UID: \"0048e137-52d0-476b-845a-4c42829de1e2\") " pod="openstack-operators/infra-operator-controller-manager-5595c7d6ff-7cg24" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.899192 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6cc65c69fc-m8ffd" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.903878 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-846c4cdcb7-4s5br"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.904716 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-846c4cdcb7-4s5br" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.919122 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.921703 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz2cp\" (UniqueName: \"kubernetes.io/projected/58ec1c2d-43a6-4582-9a21-f1083777ffe5-kube-api-access-dz2cp\") pod \"heat-operator-controller-manager-66dd9d474d-qf5kd\" (UID: \"58ec1c2d-43a6-4582-9a21-f1083777ffe5\") " pod="openstack-operators/heat-operator-controller-manager-66dd9d474d-qf5kd" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.921957 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7d559dcdbd-zxgmc" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.924094 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.926463 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69czt\" (UniqueName: \"kubernetes.io/projected/f9ca1637-123c-47eb-8183-5394a0ba6f89-kube-api-access-69czt\") pod \"neutron-operator-controller-manager-6744dd545c-mvqjq\" (UID: \"f9ca1637-123c-47eb-8183-5394a0ba6f89\") " pod="openstack-operators/neutron-operator-controller-manager-6744dd545c-mvqjq" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.926505 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lndx9\" (UniqueName: \"kubernetes.io/projected/8e08ed5d-23b6-4066-a86f-3019f8cb34c8-kube-api-access-lndx9\") pod \"manila-operator-controller-manager-fbf7bbb96-745cq\" (UID: \"8e08ed5d-23b6-4066-a86f-3019f8cb34c8\") " pod="openstack-operators/manila-operator-controller-manager-fbf7bbb96-745cq" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.926539 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvlt8\" (UniqueName: \"kubernetes.io/projected/a4bbdaf9-e480-4cd1-9bf8-304cf5266949-kube-api-access-bvlt8\") pod \"mariadb-operator-controller-manager-6f5b7bcd4-pzgkp\" (UID: \"a4bbdaf9-e480-4cd1-9bf8-304cf5266949\") " pod="openstack-operators/mariadb-operator-controller-manager-6f5b7bcd4-pzgkp" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.926611 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4g69\" (UniqueName: \"kubernetes.io/projected/50f56ad0-1409-4619-9b7a-887ba6997874-kube-api-access-k4g69\") pod \"nova-operator-controller-manager-bc5c78db9-c9skl\" (UID: \"50f56ad0-1409-4619-9b7a-887ba6997874\") " pod="openstack-operators/nova-operator-controller-manager-bc5c78db9-c9skl" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.926637 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvtr5\" (UniqueName: \"kubernetes.io/projected/c31077f0-de38-4de7-8421-a5ea7ac3ab84-kube-api-access-gvtr5\") pod \"keystone-operator-controller-manager-76b87776c9-rcz2g\" (UID: \"c31077f0-de38-4de7-8421-a5ea7ac3ab84\") " pod="openstack-operators/keystone-operator-controller-manager-76b87776c9-rcz2g" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.926656 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xgsh\" (UniqueName: \"kubernetes.io/projected/828a9554-fa0a-4458-9d63-b9e6bd6bde49-kube-api-access-2xgsh\") pod \"octavia-operator-controller-manager-56f74467c6-fzv5q\" (UID: \"828a9554-fa0a-4458-9d63-b9e6bd6bde49\") " pod="openstack-operators/octavia-operator-controller-manager-56f74467c6-fzv5q" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.927670 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-5ckpt" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.927853 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-xsqb7" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.927956 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.928159 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-846c4cdcb7-4s5br"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.946988 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.947045 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-867f54bc44-n94f2"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.948132 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-867f54bc44-n94f2" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.951126 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-659fb58c6b-m5bx9"] Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.951986 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-659fb58c6b-m5bx9" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.955485 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69czt\" (UniqueName: \"kubernetes.io/projected/f9ca1637-123c-47eb-8183-5394a0ba6f89-kube-api-access-69czt\") pod \"neutron-operator-controller-manager-6744dd545c-mvqjq\" (UID: \"f9ca1637-123c-47eb-8183-5394a0ba6f89\") " pod="openstack-operators/neutron-operator-controller-manager-6744dd545c-mvqjq" Mar 21 08:44:11 crc kubenswrapper[4696]: I0321 08:44:11.995856 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6744dd545c-mvqjq" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.000168 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-9q95l" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.004407 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-66dd9d474d-qf5kd" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.006688 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-64dc66d669-8685h" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.009327 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-srw8f" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.012406 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvtr5\" (UniqueName: \"kubernetes.io/projected/c31077f0-de38-4de7-8421-a5ea7ac3ab84-kube-api-access-gvtr5\") pod \"keystone-operator-controller-manager-76b87776c9-rcz2g\" (UID: \"c31077f0-de38-4de7-8421-a5ea7ac3ab84\") " pod="openstack-operators/keystone-operator-controller-manager-76b87776c9-rcz2g" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.025526 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lndx9\" (UniqueName: \"kubernetes.io/projected/8e08ed5d-23b6-4066-a86f-3019f8cb34c8-kube-api-access-lndx9\") pod \"manila-operator-controller-manager-fbf7bbb96-745cq\" (UID: \"8e08ed5d-23b6-4066-a86f-3019f8cb34c8\") " pod="openstack-operators/manila-operator-controller-manager-fbf7bbb96-745cq" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.028479 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsglf\" (UniqueName: \"kubernetes.io/projected/5fe44a3f-65e2-481b-9d1d-f3b3c24944db-kube-api-access-bsglf\") pod \"swift-operator-controller-manager-867f54bc44-n94f2\" (UID: \"5fe44a3f-65e2-481b-9d1d-f3b3c24944db\") " pod="openstack-operators/swift-operator-controller-manager-867f54bc44-n94f2" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.028516 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgmcs\" (UniqueName: \"kubernetes.io/projected/d2672ea4-e1e0-4809-b4e9-04e2cece2179-kube-api-access-tgmcs\") pod \"placement-operator-controller-manager-659fb58c6b-m5bx9\" (UID: \"d2672ea4-e1e0-4809-b4e9-04e2cece2179\") " pod="openstack-operators/placement-operator-controller-manager-659fb58c6b-m5bx9" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.028564 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd4sh\" (UniqueName: \"kubernetes.io/projected/026783e3-cff7-40bf-9f26-2d5e8b6006e7-kube-api-access-bd4sh\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89\" (UID: \"026783e3-cff7-40bf-9f26-2d5e8b6006e7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.028614 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvxgq\" (UniqueName: \"kubernetes.io/projected/1c77ab60-1377-4880-80a9-eb1af2009bca-kube-api-access-kvxgq\") pod \"ovn-operator-controller-manager-846c4cdcb7-4s5br\" (UID: \"1c77ab60-1377-4880-80a9-eb1af2009bca\") " pod="openstack-operators/ovn-operator-controller-manager-846c4cdcb7-4s5br" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.028673 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4g69\" (UniqueName: \"kubernetes.io/projected/50f56ad0-1409-4619-9b7a-887ba6997874-kube-api-access-k4g69\") pod \"nova-operator-controller-manager-bc5c78db9-c9skl\" (UID: \"50f56ad0-1409-4619-9b7a-887ba6997874\") " pod="openstack-operators/nova-operator-controller-manager-bc5c78db9-c9skl" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.028694 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/026783e3-cff7-40bf-9f26-2d5e8b6006e7-cert\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89\" (UID: \"026783e3-cff7-40bf-9f26-2d5e8b6006e7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.028723 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xgsh\" (UniqueName: \"kubernetes.io/projected/828a9554-fa0a-4458-9d63-b9e6bd6bde49-kube-api-access-2xgsh\") pod \"octavia-operator-controller-manager-56f74467c6-fzv5q\" (UID: \"828a9554-fa0a-4458-9d63-b9e6bd6bde49\") " pod="openstack-operators/octavia-operator-controller-manager-56f74467c6-fzv5q" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.029594 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvlt8\" (UniqueName: \"kubernetes.io/projected/a4bbdaf9-e480-4cd1-9bf8-304cf5266949-kube-api-access-bvlt8\") pod \"mariadb-operator-controller-manager-6f5b7bcd4-pzgkp\" (UID: \"a4bbdaf9-e480-4cd1-9bf8-304cf5266949\") " pod="openstack-operators/mariadb-operator-controller-manager-6f5b7bcd4-pzgkp" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.032528 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6b77b7676d-vhnjl" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.052728 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xgsh\" (UniqueName: \"kubernetes.io/projected/828a9554-fa0a-4458-9d63-b9e6bd6bde49-kube-api-access-2xgsh\") pod \"octavia-operator-controller-manager-56f74467c6-fzv5q\" (UID: \"828a9554-fa0a-4458-9d63-b9e6bd6bde49\") " pod="openstack-operators/octavia-operator-controller-manager-56f74467c6-fzv5q" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.055342 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4g69\" (UniqueName: \"kubernetes.io/projected/50f56ad0-1409-4619-9b7a-887ba6997874-kube-api-access-k4g69\") pod \"nova-operator-controller-manager-bc5c78db9-c9skl\" (UID: \"50f56ad0-1409-4619-9b7a-887ba6997874\") " pod="openstack-operators/nova-operator-controller-manager-bc5c78db9-c9skl" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.056020 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-76b87776c9-rcz2g" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.097999 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-69d78c9949-5whvx"] Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.099831 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-69d78c9949-5whvx" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.115709 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-kmdkc" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.133592 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsglf\" (UniqueName: \"kubernetes.io/projected/5fe44a3f-65e2-481b-9d1d-f3b3c24944db-kube-api-access-bsglf\") pod \"swift-operator-controller-manager-867f54bc44-n94f2\" (UID: \"5fe44a3f-65e2-481b-9d1d-f3b3c24944db\") " pod="openstack-operators/swift-operator-controller-manager-867f54bc44-n94f2" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.137258 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgmcs\" (UniqueName: \"kubernetes.io/projected/d2672ea4-e1e0-4809-b4e9-04e2cece2179-kube-api-access-tgmcs\") pod \"placement-operator-controller-manager-659fb58c6b-m5bx9\" (UID: \"d2672ea4-e1e0-4809-b4e9-04e2cece2179\") " pod="openstack-operators/placement-operator-controller-manager-659fb58c6b-m5bx9" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.137335 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd4sh\" (UniqueName: \"kubernetes.io/projected/026783e3-cff7-40bf-9f26-2d5e8b6006e7-kube-api-access-bd4sh\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89\" (UID: \"026783e3-cff7-40bf-9f26-2d5e8b6006e7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.137421 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvxgq\" (UniqueName: \"kubernetes.io/projected/1c77ab60-1377-4880-80a9-eb1af2009bca-kube-api-access-kvxgq\") pod \"ovn-operator-controller-manager-846c4cdcb7-4s5br\" (UID: \"1c77ab60-1377-4880-80a9-eb1af2009bca\") " pod="openstack-operators/ovn-operator-controller-manager-846c4cdcb7-4s5br" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.137524 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22wdk\" (UniqueName: \"kubernetes.io/projected/0ecf9ee3-d5c9-4677-b264-0ddae28df715-kube-api-access-22wdk\") pod \"telemetry-operator-controller-manager-69d78c9949-5whvx\" (UID: \"0ecf9ee3-d5c9-4677-b264-0ddae28df715\") " pod="openstack-operators/telemetry-operator-controller-manager-69d78c9949-5whvx" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.138379 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-fbf7bbb96-745cq" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.146191 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/026783e3-cff7-40bf-9f26-2d5e8b6006e7-cert\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89\" (UID: \"026783e3-cff7-40bf-9f26-2d5e8b6006e7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" Mar 21 08:44:12 crc kubenswrapper[4696]: E0321 08:44:12.147151 4696 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 08:44:12 crc kubenswrapper[4696]: E0321 08:44:12.147206 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/026783e3-cff7-40bf-9f26-2d5e8b6006e7-cert podName:026783e3-cff7-40bf-9f26-2d5e8b6006e7 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:12.647193036 +0000 UTC m=+986.768073749 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/026783e3-cff7-40bf-9f26-2d5e8b6006e7-cert") pod "openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" (UID: "026783e3-cff7-40bf-9f26-2d5e8b6006e7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.177704 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgmcs\" (UniqueName: \"kubernetes.io/projected/d2672ea4-e1e0-4809-b4e9-04e2cece2179-kube-api-access-tgmcs\") pod \"placement-operator-controller-manager-659fb58c6b-m5bx9\" (UID: \"d2672ea4-e1e0-4809-b4e9-04e2cece2179\") " pod="openstack-operators/placement-operator-controller-manager-659fb58c6b-m5bx9" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.189804 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd4sh\" (UniqueName: \"kubernetes.io/projected/026783e3-cff7-40bf-9f26-2d5e8b6006e7-kube-api-access-bd4sh\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89\" (UID: \"026783e3-cff7-40bf-9f26-2d5e8b6006e7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.190795 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsglf\" (UniqueName: \"kubernetes.io/projected/5fe44a3f-65e2-481b-9d1d-f3b3c24944db-kube-api-access-bsglf\") pod \"swift-operator-controller-manager-867f54bc44-n94f2\" (UID: \"5fe44a3f-65e2-481b-9d1d-f3b3c24944db\") " pod="openstack-operators/swift-operator-controller-manager-867f54bc44-n94f2" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.191242 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-867f54bc44-n94f2"] Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.191501 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvxgq\" (UniqueName: \"kubernetes.io/projected/1c77ab60-1377-4880-80a9-eb1af2009bca-kube-api-access-kvxgq\") pod \"ovn-operator-controller-manager-846c4cdcb7-4s5br\" (UID: \"1c77ab60-1377-4880-80a9-eb1af2009bca\") " pod="openstack-operators/ovn-operator-controller-manager-846c4cdcb7-4s5br" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.215504 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-659fb58c6b-m5bx9"] Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.224043 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-69d78c9949-5whvx"] Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.228409 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6f5b7bcd4-pzgkp" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.261646 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22wdk\" (UniqueName: \"kubernetes.io/projected/0ecf9ee3-d5c9-4677-b264-0ddae28df715-kube-api-access-22wdk\") pod \"telemetry-operator-controller-manager-69d78c9949-5whvx\" (UID: \"0ecf9ee3-d5c9-4677-b264-0ddae28df715\") " pod="openstack-operators/telemetry-operator-controller-manager-69d78c9949-5whvx" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.266904 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-8467ccb4c8-zk58x"] Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.274507 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-8467ccb4c8-zk58x" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.276807 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-89r4j" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.278303 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-8467ccb4c8-zk58x"] Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.310921 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-bc5c78db9-c9skl" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.316228 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22wdk\" (UniqueName: \"kubernetes.io/projected/0ecf9ee3-d5c9-4677-b264-0ddae28df715-kube-api-access-22wdk\") pod \"telemetry-operator-controller-manager-69d78c9949-5whvx\" (UID: \"0ecf9ee3-d5c9-4677-b264-0ddae28df715\") " pod="openstack-operators/telemetry-operator-controller-manager-69d78c9949-5whvx" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.321411 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-56f74467c6-fzv5q" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.330862 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-74d6f7b5c-5x9b6"] Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.332048 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-74d6f7b5c-5x9b6" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.342729 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-vzt76" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.349127 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-846c4cdcb7-4s5br" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.361957 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-74d6f7b5c-5x9b6"] Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.363678 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zgvf\" (UniqueName: \"kubernetes.io/projected/46974657-c404-4081-a4ef-5fc99237a02f-kube-api-access-9zgvf\") pod \"test-operator-controller-manager-8467ccb4c8-zk58x\" (UID: \"46974657-c404-4081-a4ef-5fc99237a02f\") " pod="openstack-operators/test-operator-controller-manager-8467ccb4c8-zk58x" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.363758 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0048e137-52d0-476b-845a-4c42829de1e2-cert\") pod \"infra-operator-controller-manager-5595c7d6ff-7cg24\" (UID: \"0048e137-52d0-476b-845a-4c42829de1e2\") " pod="openstack-operators/infra-operator-controller-manager-5595c7d6ff-7cg24" Mar 21 08:44:12 crc kubenswrapper[4696]: E0321 08:44:12.363949 4696 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 21 08:44:12 crc kubenswrapper[4696]: E0321 08:44:12.363997 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0048e137-52d0-476b-845a-4c42829de1e2-cert podName:0048e137-52d0-476b-845a-4c42829de1e2 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:13.363983234 +0000 UTC m=+987.484863947 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0048e137-52d0-476b-845a-4c42829de1e2-cert") pod "infra-operator-controller-manager-5595c7d6ff-7cg24" (UID: "0048e137-52d0-476b-845a-4c42829de1e2") : secret "infra-operator-webhook-server-cert" not found Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.400937 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc"] Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.402272 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.407872 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-dzw6n" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.408931 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.410644 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.410664 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc"] Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.411210 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-867f54bc44-n94f2" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.442744 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvljx"] Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.444162 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvljx" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.447261 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-clbzr" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.447698 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-659fb58c6b-m5bx9" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.465083 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-metrics-certs\") pod \"openstack-operator-controller-manager-b7fdfc6b-hrhxc\" (UID: \"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0\") " pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.465155 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zgvf\" (UniqueName: \"kubernetes.io/projected/46974657-c404-4081-a4ef-5fc99237a02f-kube-api-access-9zgvf\") pod \"test-operator-controller-manager-8467ccb4c8-zk58x\" (UID: \"46974657-c404-4081-a4ef-5fc99237a02f\") " pod="openstack-operators/test-operator-controller-manager-8467ccb4c8-zk58x" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.465182 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txlfl\" (UniqueName: \"kubernetes.io/projected/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-kube-api-access-txlfl\") pod \"openstack-operator-controller-manager-b7fdfc6b-hrhxc\" (UID: \"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0\") " pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.465204 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpxrs\" (UniqueName: \"kubernetes.io/projected/f593acbc-0a73-41ed-a949-02a3556dfc70-kube-api-access-wpxrs\") pod \"watcher-operator-controller-manager-74d6f7b5c-5x9b6\" (UID: \"f593acbc-0a73-41ed-a949-02a3556dfc70\") " pod="openstack-operators/watcher-operator-controller-manager-74d6f7b5c-5x9b6" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.465236 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-webhook-certs\") pod \"openstack-operator-controller-manager-b7fdfc6b-hrhxc\" (UID: \"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0\") " pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.480605 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-69d78c9949-5whvx" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.485278 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvljx"] Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.486087 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zgvf\" (UniqueName: \"kubernetes.io/projected/46974657-c404-4081-a4ef-5fc99237a02f-kube-api-access-9zgvf\") pod \"test-operator-controller-manager-8467ccb4c8-zk58x\" (UID: \"46974657-c404-4081-a4ef-5fc99237a02f\") " pod="openstack-operators/test-operator-controller-manager-8467ccb4c8-zk58x" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.566833 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpxrs\" (UniqueName: \"kubernetes.io/projected/f593acbc-0a73-41ed-a949-02a3556dfc70-kube-api-access-wpxrs\") pod \"watcher-operator-controller-manager-74d6f7b5c-5x9b6\" (UID: \"f593acbc-0a73-41ed-a949-02a3556dfc70\") " pod="openstack-operators/watcher-operator-controller-manager-74d6f7b5c-5x9b6" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.566899 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-webhook-certs\") pod \"openstack-operator-controller-manager-b7fdfc6b-hrhxc\" (UID: \"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0\") " pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.566961 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kznq\" (UniqueName: \"kubernetes.io/projected/1ff5cfb7-37ad-4f6a-84f1-e8dbdef06084-kube-api-access-9kznq\") pod \"rabbitmq-cluster-operator-manager-668c99d594-kvljx\" (UID: \"1ff5cfb7-37ad-4f6a-84f1-e8dbdef06084\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvljx" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.567001 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-metrics-certs\") pod \"openstack-operator-controller-manager-b7fdfc6b-hrhxc\" (UID: \"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0\") " pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.567065 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txlfl\" (UniqueName: \"kubernetes.io/projected/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-kube-api-access-txlfl\") pod \"openstack-operator-controller-manager-b7fdfc6b-hrhxc\" (UID: \"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0\") " pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:12 crc kubenswrapper[4696]: E0321 08:44:12.572307 4696 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 21 08:44:12 crc kubenswrapper[4696]: E0321 08:44:12.572408 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-metrics-certs podName:fbbc65c6-7696-4795-8b4b-7d9c7e315bf0 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:13.07239036 +0000 UTC m=+987.193271073 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-metrics-certs") pod "openstack-operator-controller-manager-b7fdfc6b-hrhxc" (UID: "fbbc65c6-7696-4795-8b4b-7d9c7e315bf0") : secret "metrics-server-cert" not found Mar 21 08:44:12 crc kubenswrapper[4696]: E0321 08:44:12.567587 4696 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 21 08:44:12 crc kubenswrapper[4696]: E0321 08:44:12.572921 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-webhook-certs podName:fbbc65c6-7696-4795-8b4b-7d9c7e315bf0 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:13.072898214 +0000 UTC m=+987.193778927 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-webhook-certs") pod "openstack-operator-controller-manager-b7fdfc6b-hrhxc" (UID: "fbbc65c6-7696-4795-8b4b-7d9c7e315bf0") : secret "webhook-server-cert" not found Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.587963 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpxrs\" (UniqueName: \"kubernetes.io/projected/f593acbc-0a73-41ed-a949-02a3556dfc70-kube-api-access-wpxrs\") pod \"watcher-operator-controller-manager-74d6f7b5c-5x9b6\" (UID: \"f593acbc-0a73-41ed-a949-02a3556dfc70\") " pod="openstack-operators/watcher-operator-controller-manager-74d6f7b5c-5x9b6" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.588692 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txlfl\" (UniqueName: \"kubernetes.io/projected/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-kube-api-access-txlfl\") pod \"openstack-operator-controller-manager-b7fdfc6b-hrhxc\" (UID: \"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0\") " pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.613888 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-8467ccb4c8-zk58x" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.621349 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6d77645966-qsd27"] Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.668072 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/026783e3-cff7-40bf-9f26-2d5e8b6006e7-cert\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89\" (UID: \"026783e3-cff7-40bf-9f26-2d5e8b6006e7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.668117 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kznq\" (UniqueName: \"kubernetes.io/projected/1ff5cfb7-37ad-4f6a-84f1-e8dbdef06084-kube-api-access-9kznq\") pod \"rabbitmq-cluster-operator-manager-668c99d594-kvljx\" (UID: \"1ff5cfb7-37ad-4f6a-84f1-e8dbdef06084\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvljx" Mar 21 08:44:12 crc kubenswrapper[4696]: E0321 08:44:12.670506 4696 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 08:44:12 crc kubenswrapper[4696]: E0321 08:44:12.670544 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/026783e3-cff7-40bf-9f26-2d5e8b6006e7-cert podName:026783e3-cff7-40bf-9f26-2d5e8b6006e7 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:13.670532151 +0000 UTC m=+987.791412864 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/026783e3-cff7-40bf-9f26-2d5e8b6006e7-cert") pod "openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" (UID: "026783e3-cff7-40bf-9f26-2d5e8b6006e7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.695600 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kznq\" (UniqueName: \"kubernetes.io/projected/1ff5cfb7-37ad-4f6a-84f1-e8dbdef06084-kube-api-access-9kznq\") pod \"rabbitmq-cluster-operator-manager-668c99d594-kvljx\" (UID: \"1ff5cfb7-37ad-4f6a-84f1-e8dbdef06084\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvljx" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.736429 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-74d6f7b5c-5x9b6" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.753860 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6d77645966-qsd27" event={"ID":"3cd62d0b-d740-4bef-ac3c-993b5ee05d40","Type":"ContainerStarted","Data":"3fa91829530320821cdd2bbddec31d10daba8591d84a981a2e62b7fe22018395"} Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.786008 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvljx" Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.788087 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7d559dcdbd-zxgmc"] Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.801418 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5cfd84c587-ldj9m"] Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.990834 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6cc65c69fc-m8ffd"] Mar 21 08:44:12 crc kubenswrapper[4696]: W0321 08:44:12.994389 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9ca1637_123c_47eb_8183_5394a0ba6f89.slice/crio-cb2d449910236f845f92acb35de34a53dd980a07bf6597a8ea18b3cf5d997ea9 WatchSource:0}: Error finding container cb2d449910236f845f92acb35de34a53dd980a07bf6597a8ea18b3cf5d997ea9: Status 404 returned error can't find the container with id cb2d449910236f845f92acb35de34a53dd980a07bf6597a8ea18b3cf5d997ea9 Mar 21 08:44:12 crc kubenswrapper[4696]: I0321 08:44:12.996882 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6744dd545c-mvqjq"] Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.033785 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-76b87776c9-rcz2g"] Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.083629 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-webhook-certs\") pod \"openstack-operator-controller-manager-b7fdfc6b-hrhxc\" (UID: \"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0\") " pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.083732 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-metrics-certs\") pod \"openstack-operator-controller-manager-b7fdfc6b-hrhxc\" (UID: \"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0\") " pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.083915 4696 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.083974 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-metrics-certs podName:fbbc65c6-7696-4795-8b4b-7d9c7e315bf0 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:14.08395435 +0000 UTC m=+988.204835063 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-metrics-certs") pod "openstack-operator-controller-manager-b7fdfc6b-hrhxc" (UID: "fbbc65c6-7696-4795-8b4b-7d9c7e315bf0") : secret "metrics-server-cert" not found Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.084144 4696 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.084254 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-webhook-certs podName:fbbc65c6-7696-4795-8b4b-7d9c7e315bf0 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:14.084231227 +0000 UTC m=+988.205112050 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-webhook-certs") pod "openstack-operator-controller-manager-b7fdfc6b-hrhxc" (UID: "fbbc65c6-7696-4795-8b4b-7d9c7e315bf0") : secret "webhook-server-cert" not found Mar 21 08:44:13 crc kubenswrapper[4696]: W0321 08:44:13.189519 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e08ed5d_23b6_4066_a86f_3019f8cb34c8.slice/crio-f0039f18c2af30733cac889e79cc9f3d964cebc63556efd3d2fa1748582276c8 WatchSource:0}: Error finding container f0039f18c2af30733cac889e79cc9f3d964cebc63556efd3d2fa1748582276c8: Status 404 returned error can't find the container with id f0039f18c2af30733cac889e79cc9f3d964cebc63556efd3d2fa1748582276c8 Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.202963 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6f5b7bcd4-pzgkp"] Mar 21 08:44:13 crc kubenswrapper[4696]: W0321 08:44:13.203271 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58ec1c2d_43a6_4582_9a21_f1083777ffe5.slice/crio-7e28803ebeb33c0838585fd9a31edd1da6e254d3b8344ed386eef3435249226b WatchSource:0}: Error finding container 7e28803ebeb33c0838585fd9a31edd1da6e254d3b8344ed386eef3435249226b: Status 404 returned error can't find the container with id 7e28803ebeb33c0838585fd9a31edd1da6e254d3b8344ed386eef3435249226b Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.227579 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-fbf7bbb96-745cq"] Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.241698 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-66dd9d474d-qf5kd"] Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.247368 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-846c4cdcb7-4s5br"] Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.261354 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-64dc66d669-8685h"] Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.268379 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6b77b7676d-vhnjl"] Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.361573 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-bc5c78db9-c9skl"] Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.375490 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-867f54bc44-n94f2"] Mar 21 08:44:13 crc kubenswrapper[4696]: W0321 08:44:13.377174 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50f56ad0_1409_4619_9b7a_887ba6997874.slice/crio-f3024a2fb440bccac0dace14392337370d159082758dc40aee9eff700aad2f5b WatchSource:0}: Error finding container f3024a2fb440bccac0dace14392337370d159082758dc40aee9eff700aad2f5b: Status 404 returned error can't find the container with id f3024a2fb440bccac0dace14392337370d159082758dc40aee9eff700aad2f5b Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.388124 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:56a4ec82efbed56683a95dd80854da49106f82b909ce3cb1eab9eaffe0e30552,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k4g69,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-bc5c78db9-c9skl_openstack-operators(50f56ad0-1409-4619-9b7a-887ba6997874): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.388756 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0048e137-52d0-476b-845a-4c42829de1e2-cert\") pod \"infra-operator-controller-manager-5595c7d6ff-7cg24\" (UID: \"0048e137-52d0-476b-845a-4c42829de1e2\") " pod="openstack-operators/infra-operator-controller-manager-5595c7d6ff-7cg24" Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.388922 4696 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.388987 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0048e137-52d0-476b-845a-4c42829de1e2-cert podName:0048e137-52d0-476b-845a-4c42829de1e2 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:15.388970015 +0000 UTC m=+989.509850728 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0048e137-52d0-476b-845a-4c42829de1e2-cert") pod "infra-operator-controller-manager-5595c7d6ff-7cg24" (UID: "0048e137-52d0-476b-845a-4c42829de1e2") : secret "infra-operator-webhook-server-cert" not found Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.390153 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-bc5c78db9-c9skl" podUID="50f56ad0-1409-4619-9b7a-887ba6997874" Mar 21 08:44:13 crc kubenswrapper[4696]: W0321 08:44:13.393033 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2672ea4_e1e0_4809_b4e9_04e2cece2179.slice/crio-ad97a91abff08f78d4595c95af426a6f6ad9230e7056f5cc33ddfcf54ef1e31e WatchSource:0}: Error finding container ad97a91abff08f78d4595c95af426a6f6ad9230e7056f5cc33ddfcf54ef1e31e: Status 404 returned error can't find the container with id ad97a91abff08f78d4595c95af426a6f6ad9230e7056f5cc33ddfcf54ef1e31e Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.394865 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-659fb58c6b-m5bx9"] Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.403363 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:6c837f09c0f3246b28931fcd0758f667ca596999558d025e06fc7b7611edec1a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tgmcs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-659fb58c6b-m5bx9_openstack-operators(d2672ea4-e1e0-4809-b4e9-04e2cece2179): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.404765 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-659fb58c6b-m5bx9" podUID="d2672ea4-e1e0-4809-b4e9-04e2cece2179" Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.487921 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-74d6f7b5c-5x9b6"] Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.495594 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-56f74467c6-fzv5q"] Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.507318 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-8467ccb4c8-zk58x"] Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.509971 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:b0ba0389a96140174eaad4ad8cc3e98118472d640bdca18046877e973f009ff4,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2xgsh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-56f74467c6-fzv5q_openstack-operators(828a9554-fa0a-4458-9d63-b9e6bd6bde49): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.510800 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:05a6fd95f5a1472c74e40b3efe58ff423cc2a00e745eea6dea19f622ef2c0953,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wpxrs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-74d6f7b5c-5x9b6_openstack-operators(f593acbc-0a73-41ed-a949-02a3556dfc70): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.511154 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-56f74467c6-fzv5q" podUID="828a9554-fa0a-4458-9d63-b9e6bd6bde49" Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.512938 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-74d6f7b5c-5x9b6" podUID="f593acbc-0a73-41ed-a949-02a3556dfc70" Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.516990 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvljx"] Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.528672 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-69d78c9949-5whvx"] Mar 21 08:44:13 crc kubenswrapper[4696]: W0321 08:44:13.547762 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ff5cfb7_37ad_4f6a_84f1_e8dbdef06084.slice/crio-71466a56ef5f56cc1a86f67834e3dcbe91ce02fca754970fe66994f786c43289 WatchSource:0}: Error finding container 71466a56ef5f56cc1a86f67834e3dcbe91ce02fca754970fe66994f786c43289: Status 404 returned error can't find the container with id 71466a56ef5f56cc1a86f67834e3dcbe91ce02fca754970fe66994f786c43289 Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.551366 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9kznq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-kvljx_openstack-operators(1ff5cfb7-37ad-4f6a-84f1-e8dbdef06084): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.552483 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvljx" podUID="1ff5cfb7-37ad-4f6a-84f1-e8dbdef06084" Mar 21 08:44:13 crc kubenswrapper[4696]: W0321 08:44:13.554964 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ecf9ee3_d5c9_4677_b264_0ddae28df715.slice/crio-ead824ef77b5cfc474c588a6d84f258ce23fb40ca4e5e5864341ecbd247f4a65 WatchSource:0}: Error finding container ead824ef77b5cfc474c588a6d84f258ce23fb40ca4e5e5864341ecbd247f4a65: Status 404 returned error can't find the container with id ead824ef77b5cfc474c588a6d84f258ce23fb40ca4e5e5864341ecbd247f4a65 Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.557004 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.177:5001/openstack-k8s-operators/telemetry-operator:ec466f39d92d34b9ce1e665f9a0fbfe564d3f1c7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-22wdk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-69d78c9949-5whvx_openstack-operators(0ecf9ee3-d5c9-4677-b264-0ddae28df715): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.558242 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-69d78c9949-5whvx" podUID="0ecf9ee3-d5c9-4677-b264-0ddae28df715" Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.692904 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/026783e3-cff7-40bf-9f26-2d5e8b6006e7-cert\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89\" (UID: \"026783e3-cff7-40bf-9f26-2d5e8b6006e7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.693113 4696 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.693471 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/026783e3-cff7-40bf-9f26-2d5e8b6006e7-cert podName:026783e3-cff7-40bf-9f26-2d5e8b6006e7 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:15.693446114 +0000 UTC m=+989.814326827 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/026783e3-cff7-40bf-9f26-2d5e8b6006e7-cert") pod "openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" (UID: "026783e3-cff7-40bf-9f26-2d5e8b6006e7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.772202 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6f5b7bcd4-pzgkp" event={"ID":"a4bbdaf9-e480-4cd1-9bf8-304cf5266949","Type":"ContainerStarted","Data":"a780c3389130e1607bba305500a260f09a58a50c4f9debf9983805595fc530de"} Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.775174 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-659fb58c6b-m5bx9" event={"ID":"d2672ea4-e1e0-4809-b4e9-04e2cece2179","Type":"ContainerStarted","Data":"ad97a91abff08f78d4595c95af426a6f6ad9230e7056f5cc33ddfcf54ef1e31e"} Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.776422 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:6c837f09c0f3246b28931fcd0758f667ca596999558d025e06fc7b7611edec1a\\\"\"" pod="openstack-operators/placement-operator-controller-manager-659fb58c6b-m5bx9" podUID="d2672ea4-e1e0-4809-b4e9-04e2cece2179" Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.777093 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvljx" event={"ID":"1ff5cfb7-37ad-4f6a-84f1-e8dbdef06084","Type":"ContainerStarted","Data":"71466a56ef5f56cc1a86f67834e3dcbe91ce02fca754970fe66994f786c43289"} Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.778443 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvljx" podUID="1ff5cfb7-37ad-4f6a-84f1-e8dbdef06084" Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.779343 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-74d6f7b5c-5x9b6" event={"ID":"f593acbc-0a73-41ed-a949-02a3556dfc70","Type":"ContainerStarted","Data":"9755aac7d4d916a301bead3b051ce04de6fe7707473773b6a72a67fcbfb92d80"} Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.780491 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:05a6fd95f5a1472c74e40b3efe58ff423cc2a00e745eea6dea19f622ef2c0953\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-74d6f7b5c-5x9b6" podUID="f593acbc-0a73-41ed-a949-02a3556dfc70" Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.781142 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-66dd9d474d-qf5kd" event={"ID":"58ec1c2d-43a6-4582-9a21-f1083777ffe5","Type":"ContainerStarted","Data":"7e28803ebeb33c0838585fd9a31edd1da6e254d3b8344ed386eef3435249226b"} Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.783887 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6cc65c69fc-m8ffd" event={"ID":"530eca0b-da29-4883-ad3a-1faa1563c65d","Type":"ContainerStarted","Data":"97e6bef9b7d2560ff76b85af2dc7882ef7b9cc1e54534266aaac0a1b5d21f25a"} Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.786566 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5cfd84c587-ldj9m" event={"ID":"53a9e616-007d-4c0c-8b0a-799d7188d9ab","Type":"ContainerStarted","Data":"3d37aecc6cfe43534a8ccf39bcdf93e9a42dd88ade8ac007888a378421ed4ace"} Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.790524 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6b77b7676d-vhnjl" event={"ID":"74fc6c6a-6eb4-4459-aed9-62073e0472e4","Type":"ContainerStarted","Data":"a26a85cedbe39a433ce7cbdb2a2e82e2688830978cc91f5bc97e52b27d1cb0b4"} Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.793561 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-64dc66d669-8685h" event={"ID":"9b05c8a9-af90-4497-94f3-8fed890ab3b0","Type":"ContainerStarted","Data":"ee99604f86f318cd1d9625f776ca7c8e6cef17505e3ef887cdc56ab822782261"} Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.798305 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-69d78c9949-5whvx" event={"ID":"0ecf9ee3-d5c9-4677-b264-0ddae28df715","Type":"ContainerStarted","Data":"ead824ef77b5cfc474c588a6d84f258ce23fb40ca4e5e5864341ecbd247f4a65"} Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.799418 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.177:5001/openstack-k8s-operators/telemetry-operator:ec466f39d92d34b9ce1e665f9a0fbfe564d3f1c7\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-69d78c9949-5whvx" podUID="0ecf9ee3-d5c9-4677-b264-0ddae28df715" Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.801010 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-bc5c78db9-c9skl" event={"ID":"50f56ad0-1409-4619-9b7a-887ba6997874","Type":"ContainerStarted","Data":"f3024a2fb440bccac0dace14392337370d159082758dc40aee9eff700aad2f5b"} Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.802754 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8467ccb4c8-zk58x" event={"ID":"46974657-c404-4081-a4ef-5fc99237a02f","Type":"ContainerStarted","Data":"620db30ecf20ec9091052f9259e58b502fc03ef8977239a6c542b5175ef1d0f3"} Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.803564 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:56a4ec82efbed56683a95dd80854da49106f82b909ce3cb1eab9eaffe0e30552\\\"\"" pod="openstack-operators/nova-operator-controller-manager-bc5c78db9-c9skl" podUID="50f56ad0-1409-4619-9b7a-887ba6997874" Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.806262 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-56f74467c6-fzv5q" event={"ID":"828a9554-fa0a-4458-9d63-b9e6bd6bde49","Type":"ContainerStarted","Data":"cf79fa0c6f92b0d3c7189725095b60bba0c0c6ad92fcd83cad2ac43f6fc56de2"} Mar 21 08:44:13 crc kubenswrapper[4696]: E0321 08:44:13.807532 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:b0ba0389a96140174eaad4ad8cc3e98118472d640bdca18046877e973f009ff4\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-56f74467c6-fzv5q" podUID="828a9554-fa0a-4458-9d63-b9e6bd6bde49" Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.812758 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-76b87776c9-rcz2g" event={"ID":"c31077f0-de38-4de7-8421-a5ea7ac3ab84","Type":"ContainerStarted","Data":"8b408b45633baa280e1b644ebd2bba806f701ca99fe375d1ba11aa8812f12b98"} Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.819896 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-fbf7bbb96-745cq" event={"ID":"8e08ed5d-23b6-4066-a86f-3019f8cb34c8","Type":"ContainerStarted","Data":"f0039f18c2af30733cac889e79cc9f3d964cebc63556efd3d2fa1748582276c8"} Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.825479 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7d559dcdbd-zxgmc" event={"ID":"9ce080a1-65e5-4235-b181-448e8987c590","Type":"ContainerStarted","Data":"776170586ab09a42d902256ff9689d11b57349e35ded38e49ea96998e036944e"} Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.826896 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6744dd545c-mvqjq" event={"ID":"f9ca1637-123c-47eb-8183-5394a0ba6f89","Type":"ContainerStarted","Data":"cb2d449910236f845f92acb35de34a53dd980a07bf6597a8ea18b3cf5d997ea9"} Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.828156 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-867f54bc44-n94f2" event={"ID":"5fe44a3f-65e2-481b-9d1d-f3b3c24944db","Type":"ContainerStarted","Data":"e5c58d12ea933fcd74f7367561e47acbe0f7bca6bdb90a38bb51010c359542a2"} Mar 21 08:44:13 crc kubenswrapper[4696]: I0321 08:44:13.830060 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-846c4cdcb7-4s5br" event={"ID":"1c77ab60-1377-4880-80a9-eb1af2009bca","Type":"ContainerStarted","Data":"7ad05cb33ab11e8d4eeb15d40e4a7c8892e6135f4aa0317904027560c3e4aaa7"} Mar 21 08:44:14 crc kubenswrapper[4696]: I0321 08:44:14.109518 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-metrics-certs\") pod \"openstack-operator-controller-manager-b7fdfc6b-hrhxc\" (UID: \"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0\") " pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:14 crc kubenswrapper[4696]: I0321 08:44:14.109638 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-webhook-certs\") pod \"openstack-operator-controller-manager-b7fdfc6b-hrhxc\" (UID: \"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0\") " pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:14 crc kubenswrapper[4696]: E0321 08:44:14.109782 4696 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 21 08:44:14 crc kubenswrapper[4696]: E0321 08:44:14.109872 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-webhook-certs podName:fbbc65c6-7696-4795-8b4b-7d9c7e315bf0 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:16.109858156 +0000 UTC m=+990.230738869 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-webhook-certs") pod "openstack-operator-controller-manager-b7fdfc6b-hrhxc" (UID: "fbbc65c6-7696-4795-8b4b-7d9c7e315bf0") : secret "webhook-server-cert" not found Mar 21 08:44:14 crc kubenswrapper[4696]: E0321 08:44:14.110182 4696 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 21 08:44:14 crc kubenswrapper[4696]: E0321 08:44:14.110455 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-metrics-certs podName:fbbc65c6-7696-4795-8b4b-7d9c7e315bf0 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:16.110433702 +0000 UTC m=+990.231314465 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-metrics-certs") pod "openstack-operator-controller-manager-b7fdfc6b-hrhxc" (UID: "fbbc65c6-7696-4795-8b4b-7d9c7e315bf0") : secret "metrics-server-cert" not found Mar 21 08:44:14 crc kubenswrapper[4696]: E0321 08:44:14.867074 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:05a6fd95f5a1472c74e40b3efe58ff423cc2a00e745eea6dea19f622ef2c0953\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-74d6f7b5c-5x9b6" podUID="f593acbc-0a73-41ed-a949-02a3556dfc70" Mar 21 08:44:14 crc kubenswrapper[4696]: E0321 08:44:14.867390 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:6c837f09c0f3246b28931fcd0758f667ca596999558d025e06fc7b7611edec1a\\\"\"" pod="openstack-operators/placement-operator-controller-manager-659fb58c6b-m5bx9" podUID="d2672ea4-e1e0-4809-b4e9-04e2cece2179" Mar 21 08:44:14 crc kubenswrapper[4696]: E0321 08:44:14.867437 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvljx" podUID="1ff5cfb7-37ad-4f6a-84f1-e8dbdef06084" Mar 21 08:44:14 crc kubenswrapper[4696]: E0321 08:44:14.867477 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.177:5001/openstack-k8s-operators/telemetry-operator:ec466f39d92d34b9ce1e665f9a0fbfe564d3f1c7\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-69d78c9949-5whvx" podUID="0ecf9ee3-d5c9-4677-b264-0ddae28df715" Mar 21 08:44:14 crc kubenswrapper[4696]: E0321 08:44:14.867665 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:56a4ec82efbed56683a95dd80854da49106f82b909ce3cb1eab9eaffe0e30552\\\"\"" pod="openstack-operators/nova-operator-controller-manager-bc5c78db9-c9skl" podUID="50f56ad0-1409-4619-9b7a-887ba6997874" Mar 21 08:44:14 crc kubenswrapper[4696]: E0321 08:44:14.875976 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:b0ba0389a96140174eaad4ad8cc3e98118472d640bdca18046877e973f009ff4\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-56f74467c6-fzv5q" podUID="828a9554-fa0a-4458-9d63-b9e6bd6bde49" Mar 21 08:44:15 crc kubenswrapper[4696]: I0321 08:44:15.432543 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0048e137-52d0-476b-845a-4c42829de1e2-cert\") pod \"infra-operator-controller-manager-5595c7d6ff-7cg24\" (UID: \"0048e137-52d0-476b-845a-4c42829de1e2\") " pod="openstack-operators/infra-operator-controller-manager-5595c7d6ff-7cg24" Mar 21 08:44:15 crc kubenswrapper[4696]: E0321 08:44:15.432873 4696 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 21 08:44:15 crc kubenswrapper[4696]: E0321 08:44:15.432954 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0048e137-52d0-476b-845a-4c42829de1e2-cert podName:0048e137-52d0-476b-845a-4c42829de1e2 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:19.43293102 +0000 UTC m=+993.553811723 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0048e137-52d0-476b-845a-4c42829de1e2-cert") pod "infra-operator-controller-manager-5595c7d6ff-7cg24" (UID: "0048e137-52d0-476b-845a-4c42829de1e2") : secret "infra-operator-webhook-server-cert" not found Mar 21 08:44:15 crc kubenswrapper[4696]: I0321 08:44:15.748796 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/026783e3-cff7-40bf-9f26-2d5e8b6006e7-cert\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89\" (UID: \"026783e3-cff7-40bf-9f26-2d5e8b6006e7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" Mar 21 08:44:15 crc kubenswrapper[4696]: E0321 08:44:15.749502 4696 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 08:44:15 crc kubenswrapper[4696]: E0321 08:44:15.749656 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/026783e3-cff7-40bf-9f26-2d5e8b6006e7-cert podName:026783e3-cff7-40bf-9f26-2d5e8b6006e7 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:19.749585076 +0000 UTC m=+993.870465789 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/026783e3-cff7-40bf-9f26-2d5e8b6006e7-cert") pod "openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" (UID: "026783e3-cff7-40bf-9f26-2d5e8b6006e7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 08:44:16 crc kubenswrapper[4696]: I0321 08:44:16.154210 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-metrics-certs\") pod \"openstack-operator-controller-manager-b7fdfc6b-hrhxc\" (UID: \"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0\") " pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:16 crc kubenswrapper[4696]: I0321 08:44:16.154314 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-webhook-certs\") pod \"openstack-operator-controller-manager-b7fdfc6b-hrhxc\" (UID: \"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0\") " pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:16 crc kubenswrapper[4696]: E0321 08:44:16.154438 4696 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 21 08:44:16 crc kubenswrapper[4696]: E0321 08:44:16.154450 4696 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 21 08:44:16 crc kubenswrapper[4696]: E0321 08:44:16.154497 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-webhook-certs podName:fbbc65c6-7696-4795-8b4b-7d9c7e315bf0 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:20.15448191 +0000 UTC m=+994.275362613 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-webhook-certs") pod "openstack-operator-controller-manager-b7fdfc6b-hrhxc" (UID: "fbbc65c6-7696-4795-8b4b-7d9c7e315bf0") : secret "webhook-server-cert" not found Mar 21 08:44:16 crc kubenswrapper[4696]: E0321 08:44:16.154685 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-metrics-certs podName:fbbc65c6-7696-4795-8b4b-7d9c7e315bf0 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:20.154646044 +0000 UTC m=+994.275526757 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-metrics-certs") pod "openstack-operator-controller-manager-b7fdfc6b-hrhxc" (UID: "fbbc65c6-7696-4795-8b4b-7d9c7e315bf0") : secret "metrics-server-cert" not found Mar 21 08:44:19 crc kubenswrapper[4696]: I0321 08:44:19.522382 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0048e137-52d0-476b-845a-4c42829de1e2-cert\") pod \"infra-operator-controller-manager-5595c7d6ff-7cg24\" (UID: \"0048e137-52d0-476b-845a-4c42829de1e2\") " pod="openstack-operators/infra-operator-controller-manager-5595c7d6ff-7cg24" Mar 21 08:44:19 crc kubenswrapper[4696]: E0321 08:44:19.522655 4696 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 21 08:44:19 crc kubenswrapper[4696]: E0321 08:44:19.522950 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0048e137-52d0-476b-845a-4c42829de1e2-cert podName:0048e137-52d0-476b-845a-4c42829de1e2 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:27.522918237 +0000 UTC m=+1001.643798990 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0048e137-52d0-476b-845a-4c42829de1e2-cert") pod "infra-operator-controller-manager-5595c7d6ff-7cg24" (UID: "0048e137-52d0-476b-845a-4c42829de1e2") : secret "infra-operator-webhook-server-cert" not found Mar 21 08:44:19 crc kubenswrapper[4696]: I0321 08:44:19.828651 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/026783e3-cff7-40bf-9f26-2d5e8b6006e7-cert\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89\" (UID: \"026783e3-cff7-40bf-9f26-2d5e8b6006e7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" Mar 21 08:44:19 crc kubenswrapper[4696]: E0321 08:44:19.828939 4696 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 08:44:19 crc kubenswrapper[4696]: E0321 08:44:19.829034 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/026783e3-cff7-40bf-9f26-2d5e8b6006e7-cert podName:026783e3-cff7-40bf-9f26-2d5e8b6006e7 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:27.829006921 +0000 UTC m=+1001.949887714 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/026783e3-cff7-40bf-9f26-2d5e8b6006e7-cert") pod "openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" (UID: "026783e3-cff7-40bf-9f26-2d5e8b6006e7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 21 08:44:20 crc kubenswrapper[4696]: I0321 08:44:20.240805 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-metrics-certs\") pod \"openstack-operator-controller-manager-b7fdfc6b-hrhxc\" (UID: \"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0\") " pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:20 crc kubenswrapper[4696]: I0321 08:44:20.240941 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-webhook-certs\") pod \"openstack-operator-controller-manager-b7fdfc6b-hrhxc\" (UID: \"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0\") " pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:20 crc kubenswrapper[4696]: E0321 08:44:20.240968 4696 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 21 08:44:20 crc kubenswrapper[4696]: E0321 08:44:20.241045 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-metrics-certs podName:fbbc65c6-7696-4795-8b4b-7d9c7e315bf0 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:28.241027662 +0000 UTC m=+1002.361908375 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-metrics-certs") pod "openstack-operator-controller-manager-b7fdfc6b-hrhxc" (UID: "fbbc65c6-7696-4795-8b4b-7d9c7e315bf0") : secret "metrics-server-cert" not found Mar 21 08:44:20 crc kubenswrapper[4696]: E0321 08:44:20.241070 4696 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 21 08:44:20 crc kubenswrapper[4696]: E0321 08:44:20.241120 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-webhook-certs podName:fbbc65c6-7696-4795-8b4b-7d9c7e315bf0 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:28.241107624 +0000 UTC m=+1002.361988337 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-webhook-certs") pod "openstack-operator-controller-manager-b7fdfc6b-hrhxc" (UID: "fbbc65c6-7696-4795-8b4b-7d9c7e315bf0") : secret "webhook-server-cert" not found Mar 21 08:44:27 crc kubenswrapper[4696]: I0321 08:44:27.550053 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0048e137-52d0-476b-845a-4c42829de1e2-cert\") pod \"infra-operator-controller-manager-5595c7d6ff-7cg24\" (UID: \"0048e137-52d0-476b-845a-4c42829de1e2\") " pod="openstack-operators/infra-operator-controller-manager-5595c7d6ff-7cg24" Mar 21 08:44:27 crc kubenswrapper[4696]: I0321 08:44:27.555427 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0048e137-52d0-476b-845a-4c42829de1e2-cert\") pod \"infra-operator-controller-manager-5595c7d6ff-7cg24\" (UID: \"0048e137-52d0-476b-845a-4c42829de1e2\") " pod="openstack-operators/infra-operator-controller-manager-5595c7d6ff-7cg24" Mar 21 08:44:27 crc kubenswrapper[4696]: I0321 08:44:27.618716 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5595c7d6ff-7cg24" Mar 21 08:44:27 crc kubenswrapper[4696]: I0321 08:44:27.855419 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/026783e3-cff7-40bf-9f26-2d5e8b6006e7-cert\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89\" (UID: \"026783e3-cff7-40bf-9f26-2d5e8b6006e7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" Mar 21 08:44:27 crc kubenswrapper[4696]: I0321 08:44:27.864106 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/026783e3-cff7-40bf-9f26-2d5e8b6006e7-cert\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89\" (UID: \"026783e3-cff7-40bf-9f26-2d5e8b6006e7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" Mar 21 08:44:27 crc kubenswrapper[4696]: I0321 08:44:27.969716 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" Mar 21 08:44:28 crc kubenswrapper[4696]: I0321 08:44:28.260755 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-webhook-certs\") pod \"openstack-operator-controller-manager-b7fdfc6b-hrhxc\" (UID: \"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0\") " pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:28 crc kubenswrapper[4696]: I0321 08:44:28.260838 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-metrics-certs\") pod \"openstack-operator-controller-manager-b7fdfc6b-hrhxc\" (UID: \"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0\") " pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:28 crc kubenswrapper[4696]: E0321 08:44:28.260922 4696 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 21 08:44:28 crc kubenswrapper[4696]: E0321 08:44:28.260989 4696 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 21 08:44:28 crc kubenswrapper[4696]: E0321 08:44:28.260998 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-webhook-certs podName:fbbc65c6-7696-4795-8b4b-7d9c7e315bf0 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:44.260977768 +0000 UTC m=+1018.381858481 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-webhook-certs") pod "openstack-operator-controller-manager-b7fdfc6b-hrhxc" (UID: "fbbc65c6-7696-4795-8b4b-7d9c7e315bf0") : secret "webhook-server-cert" not found Mar 21 08:44:28 crc kubenswrapper[4696]: E0321 08:44:28.261038 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-metrics-certs podName:fbbc65c6-7696-4795-8b4b-7d9c7e315bf0 nodeName:}" failed. No retries permitted until 2026-03-21 08:44:44.261024459 +0000 UTC m=+1018.381905172 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-metrics-certs") pod "openstack-operator-controller-manager-b7fdfc6b-hrhxc" (UID: "fbbc65c6-7696-4795-8b4b-7d9c7e315bf0") : secret "metrics-server-cert" not found Mar 21 08:44:28 crc kubenswrapper[4696]: E0321 08:44:28.987786 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:0e0d48e3ca53577e20c81a87f0be6b3254c0b8418e3b446b68c8b5849af7213e" Mar 21 08:44:28 crc kubenswrapper[4696]: E0321 08:44:28.987995 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:0e0d48e3ca53577e20c81a87f0be6b3254c0b8418e3b446b68c8b5849af7213e,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bsglf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-867f54bc44-n94f2_openstack-operators(5fe44a3f-65e2-481b-9d1d-f3b3c24944db): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 08:44:28 crc kubenswrapper[4696]: E0321 08:44:28.989187 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-867f54bc44-n94f2" podUID="5fe44a3f-65e2-481b-9d1d-f3b3c24944db" Mar 21 08:44:29 crc kubenswrapper[4696]: E0321 08:44:29.190387 4696 log.go:32] "ImageFsInfo from image service failed" err="rpc error: code = Unknown desc = get image fs info unable to get usage for /var/lib/containers/storage/overlay-images: get disk usage for path /var/lib/containers/storage/overlay-images: lstat /var/lib/containers/storage/overlay-images/.tmp-images.json3986759512: no such file or directory" Mar 21 08:44:29 crc kubenswrapper[4696]: E0321 08:44:29.190588 4696 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get imageFs stats: missing image stats: nil" Mar 21 08:44:29 crc kubenswrapper[4696]: I0321 08:44:29.531761 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89"] Mar 21 08:44:29 crc kubenswrapper[4696]: I0321 08:44:29.677756 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5595c7d6ff-7cg24"] Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.000116 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6744dd545c-mvqjq" event={"ID":"f9ca1637-123c-47eb-8183-5394a0ba6f89","Type":"ContainerStarted","Data":"a19ba0aeb81c9ddc7b65e851fb43b242ee3a325b7fcf2c6e86f49c1187843f40"} Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.000457 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-6744dd545c-mvqjq" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.010355 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-66dd9d474d-qf5kd" event={"ID":"58ec1c2d-43a6-4582-9a21-f1083777ffe5","Type":"ContainerStarted","Data":"0f0bb05abc264288c331b3ed0b57db0b525cb3f9074c13c4089c2d9b310f0aad"} Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.010918 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-66dd9d474d-qf5kd" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.024793 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6f5b7bcd4-pzgkp" event={"ID":"a4bbdaf9-e480-4cd1-9bf8-304cf5266949","Type":"ContainerStarted","Data":"5cfd650abb33c30e84f93d744c1431d7f2d61e606651431b64bf25ab1f7e3fd2"} Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.025543 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6f5b7bcd4-pzgkp" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.033864 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-6744dd545c-mvqjq" podStartSLOduration=2.828647859 podStartE2EDuration="19.033844745s" podCreationTimestamp="2026-03-21 08:44:11 +0000 UTC" firstStartedPulling="2026-03-21 08:44:12.997570624 +0000 UTC m=+987.118451337" lastFinishedPulling="2026-03-21 08:44:29.20276751 +0000 UTC m=+1003.323648223" observedRunningTime="2026-03-21 08:44:30.033279329 +0000 UTC m=+1004.154160042" watchObservedRunningTime="2026-03-21 08:44:30.033844745 +0000 UTC m=+1004.154725458" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.039857 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6cc65c69fc-m8ffd" event={"ID":"530eca0b-da29-4883-ad3a-1faa1563c65d","Type":"ContainerStarted","Data":"03b570144d477c795eb4c9bf1f52591893979d2fc033e5ee162eb1bcae0b4a15"} Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.040229 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-6cc65c69fc-m8ffd" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.045626 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6b77b7676d-vhnjl" event={"ID":"74fc6c6a-6eb4-4459-aed9-62073e0472e4","Type":"ContainerStarted","Data":"14b23edcf9c207cad692d11c97be97e506228a7b6379c09635bffeaf160b7273"} Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.046329 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6b77b7676d-vhnjl" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.047990 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" event={"ID":"026783e3-cff7-40bf-9f26-2d5e8b6006e7","Type":"ContainerStarted","Data":"4827af65c42ed726d080287c048ada5c5a599af5de017baf6cbb42df7eb88465"} Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.064173 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-66dd9d474d-qf5kd" podStartSLOduration=3.084713232 podStartE2EDuration="19.064148993s" podCreationTimestamp="2026-03-21 08:44:11 +0000 UTC" firstStartedPulling="2026-03-21 08:44:13.224494051 +0000 UTC m=+987.345374764" lastFinishedPulling="2026-03-21 08:44:29.203929812 +0000 UTC m=+1003.324810525" observedRunningTime="2026-03-21 08:44:30.057487538 +0000 UTC m=+1004.178368261" watchObservedRunningTime="2026-03-21 08:44:30.064148993 +0000 UTC m=+1004.185029706" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.070070 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-846c4cdcb7-4s5br" event={"ID":"1c77ab60-1377-4880-80a9-eb1af2009bca","Type":"ContainerStarted","Data":"ff9cdd6877d730d4c5a85cbb5bb7820e06c2732adc5eb73d495ee706bdecf746"} Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.070801 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-846c4cdcb7-4s5br" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.083070 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-76b87776c9-rcz2g" event={"ID":"c31077f0-de38-4de7-8421-a5ea7ac3ab84","Type":"ContainerStarted","Data":"00208f3e0721cd484bc0372820fd5dcbefe9f0d0ec09e417005be9f3149661d2"} Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.083923 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-76b87776c9-rcz2g" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.084518 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6f5b7bcd4-pzgkp" podStartSLOduration=3.095467168 podStartE2EDuration="19.084503074s" podCreationTimestamp="2026-03-21 08:44:11 +0000 UTC" firstStartedPulling="2026-03-21 08:44:13.215571365 +0000 UTC m=+987.336452078" lastFinishedPulling="2026-03-21 08:44:29.204607271 +0000 UTC m=+1003.325487984" observedRunningTime="2026-03-21 08:44:30.081282836 +0000 UTC m=+1004.202163559" watchObservedRunningTime="2026-03-21 08:44:30.084503074 +0000 UTC m=+1004.205383787" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.099907 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5595c7d6ff-7cg24" event={"ID":"0048e137-52d0-476b-845a-4c42829de1e2","Type":"ContainerStarted","Data":"56e2b1cc36d20a31d1df84aae65eec26ad3beacb60d1258ee2c6b8e78ea58dae"} Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.111242 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6d77645966-qsd27" event={"ID":"3cd62d0b-d740-4bef-ac3c-993b5ee05d40","Type":"ContainerStarted","Data":"d931a44647f7ae4f16e1ed8585ce577860fb56e4ff04844faaf17b714ceea501"} Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.111467 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6d77645966-qsd27" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.115108 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6b77b7676d-vhnjl" podStartSLOduration=3.171869648 podStartE2EDuration="19.115089699s" podCreationTimestamp="2026-03-21 08:44:11 +0000 UTC" firstStartedPulling="2026-03-21 08:44:13.259176449 +0000 UTC m=+987.380057162" lastFinishedPulling="2026-03-21 08:44:29.2023965 +0000 UTC m=+1003.323277213" observedRunningTime="2026-03-21 08:44:30.111062008 +0000 UTC m=+1004.231942721" watchObservedRunningTime="2026-03-21 08:44:30.115089699 +0000 UTC m=+1004.235970412" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.129126 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5cfd84c587-ldj9m" event={"ID":"53a9e616-007d-4c0c-8b0a-799d7188d9ab","Type":"ContainerStarted","Data":"10a3c5cf49a349328f4d51de75bb859e4532995def087541122875a958fda79b"} Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.129887 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-5cfd84c587-ldj9m" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.146085 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7d559dcdbd-zxgmc" event={"ID":"9ce080a1-65e5-4235-b181-448e8987c590","Type":"ContainerStarted","Data":"8880dfc6deedc01a9288484cce6f05f632317887e1cb9eb195b25421677d061a"} Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.146375 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7d559dcdbd-zxgmc" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.151256 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-6cc65c69fc-m8ffd" podStartSLOduration=2.952320914 podStartE2EDuration="19.151235038s" podCreationTimestamp="2026-03-21 08:44:11 +0000 UTC" firstStartedPulling="2026-03-21 08:44:13.005665397 +0000 UTC m=+987.126546110" lastFinishedPulling="2026-03-21 08:44:29.204579521 +0000 UTC m=+1003.325460234" observedRunningTime="2026-03-21 08:44:30.141240482 +0000 UTC m=+1004.262121215" watchObservedRunningTime="2026-03-21 08:44:30.151235038 +0000 UTC m=+1004.272115751" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.191390 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-64dc66d669-8685h" event={"ID":"9b05c8a9-af90-4497-94f3-8fed890ab3b0","Type":"ContainerStarted","Data":"e38bbb2dc0bbe21c4647eae6af4b403bab095cea40f949863bd48048c2f2bfff"} Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.192062 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-64dc66d669-8685h" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.238675 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-846c4cdcb7-4s5br" podStartSLOduration=3.294770513 podStartE2EDuration="19.238657572s" podCreationTimestamp="2026-03-21 08:44:11 +0000 UTC" firstStartedPulling="2026-03-21 08:44:13.258975554 +0000 UTC m=+987.379856267" lastFinishedPulling="2026-03-21 08:44:29.202862613 +0000 UTC m=+1003.323743326" observedRunningTime="2026-03-21 08:44:30.186997145 +0000 UTC m=+1004.307877858" watchObservedRunningTime="2026-03-21 08:44:30.238657572 +0000 UTC m=+1004.359538285" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.260138 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-fbf7bbb96-745cq" event={"ID":"8e08ed5d-23b6-4066-a86f-3019f8cb34c8","Type":"ContainerStarted","Data":"3ddb54c876ffaf8a34a7bd6e9b9467eb1b25412a133a20bb17bb4a432194e1f1"} Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.260614 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-fbf7bbb96-745cq" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.271860 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-5cfd84c587-ldj9m" podStartSLOduration=3.012098777 podStartE2EDuration="19.271845989s" podCreationTimestamp="2026-03-21 08:44:11 +0000 UTC" firstStartedPulling="2026-03-21 08:44:12.943288575 +0000 UTC m=+987.064169288" lastFinishedPulling="2026-03-21 08:44:29.203035777 +0000 UTC m=+1003.323916500" observedRunningTime="2026-03-21 08:44:30.240094162 +0000 UTC m=+1004.360974875" watchObservedRunningTime="2026-03-21 08:44:30.271845989 +0000 UTC m=+1004.392726702" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.275677 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8467ccb4c8-zk58x" event={"ID":"46974657-c404-4081-a4ef-5fc99237a02f","Type":"ContainerStarted","Data":"b9b69e43ae4f703121353ead75ed80c576abdb8b5902d1d41e514f3a2a61edc8"} Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.275828 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-8467ccb4c8-zk58x" Mar 21 08:44:30 crc kubenswrapper[4696]: E0321 08:44:30.285540 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:0e0d48e3ca53577e20c81a87f0be6b3254c0b8418e3b446b68c8b5849af7213e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-867f54bc44-n94f2" podUID="5fe44a3f-65e2-481b-9d1d-f3b3c24944db" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.318808 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-76b87776c9-rcz2g" podStartSLOduration=3.136698396 podStartE2EDuration="19.318794805s" podCreationTimestamp="2026-03-21 08:44:11 +0000 UTC" firstStartedPulling="2026-03-21 08:44:13.022476992 +0000 UTC m=+987.143357705" lastFinishedPulling="2026-03-21 08:44:29.204573401 +0000 UTC m=+1003.325454114" observedRunningTime="2026-03-21 08:44:30.317004566 +0000 UTC m=+1004.437885279" watchObservedRunningTime="2026-03-21 08:44:30.318794805 +0000 UTC m=+1004.439675518" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.322310 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6d77645966-qsd27" podStartSLOduration=2.831083026 podStartE2EDuration="19.322302073s" podCreationTimestamp="2026-03-21 08:44:11 +0000 UTC" firstStartedPulling="2026-03-21 08:44:12.713418965 +0000 UTC m=+986.834299678" lastFinishedPulling="2026-03-21 08:44:29.204638012 +0000 UTC m=+1003.325518725" observedRunningTime="2026-03-21 08:44:30.271121809 +0000 UTC m=+1004.392002522" watchObservedRunningTime="2026-03-21 08:44:30.322302073 +0000 UTC m=+1004.443182786" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.341905 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.341956 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.341995 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.342579 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"06727ab615d47d51e86a72ca06fb67dd4f0d596dbb88f8ea56eca5339ae5cc33"} pod="openshift-machine-config-operator/machine-config-daemon-z7srw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.342627 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" containerID="cri-o://06727ab615d47d51e86a72ca06fb67dd4f0d596dbb88f8ea56eca5339ae5cc33" gracePeriod=600 Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.373988 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7d559dcdbd-zxgmc" podStartSLOduration=3.014289547 podStartE2EDuration="19.37397168s" podCreationTimestamp="2026-03-21 08:44:11 +0000 UTC" firstStartedPulling="2026-03-21 08:44:12.842787229 +0000 UTC m=+986.963667942" lastFinishedPulling="2026-03-21 08:44:29.202469362 +0000 UTC m=+1003.323350075" observedRunningTime="2026-03-21 08:44:30.372223922 +0000 UTC m=+1004.493104645" watchObservedRunningTime="2026-03-21 08:44:30.37397168 +0000 UTC m=+1004.494852393" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.432808 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-fbf7bbb96-745cq" podStartSLOduration=3.426795569 podStartE2EDuration="19.432793254s" podCreationTimestamp="2026-03-21 08:44:11 +0000 UTC" firstStartedPulling="2026-03-21 08:44:13.198619657 +0000 UTC m=+987.319500370" lastFinishedPulling="2026-03-21 08:44:29.204617342 +0000 UTC m=+1003.325498055" observedRunningTime="2026-03-21 08:44:30.398074885 +0000 UTC m=+1004.518955598" watchObservedRunningTime="2026-03-21 08:44:30.432793254 +0000 UTC m=+1004.553673957" Mar 21 08:44:30 crc kubenswrapper[4696]: I0321 08:44:30.433368 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-8467ccb4c8-zk58x" podStartSLOduration=3.75409896 podStartE2EDuration="19.43336341s" podCreationTimestamp="2026-03-21 08:44:11 +0000 UTC" firstStartedPulling="2026-03-21 08:44:13.525327161 +0000 UTC m=+987.646207874" lastFinishedPulling="2026-03-21 08:44:29.204591611 +0000 UTC m=+1003.325472324" observedRunningTime="2026-03-21 08:44:30.431032596 +0000 UTC m=+1004.551913309" watchObservedRunningTime="2026-03-21 08:44:30.43336341 +0000 UTC m=+1004.554244123" Mar 21 08:44:31 crc kubenswrapper[4696]: I0321 08:44:31.296022 4696 generic.go:334] "Generic (PLEG): container finished" podID="daaf227a-2305-495e-8495-a280abcd8e10" containerID="06727ab615d47d51e86a72ca06fb67dd4f0d596dbb88f8ea56eca5339ae5cc33" exitCode=0 Mar 21 08:44:31 crc kubenswrapper[4696]: I0321 08:44:31.296222 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerDied","Data":"06727ab615d47d51e86a72ca06fb67dd4f0d596dbb88f8ea56eca5339ae5cc33"} Mar 21 08:44:31 crc kubenswrapper[4696]: I0321 08:44:31.296989 4696 scope.go:117] "RemoveContainer" containerID="a451c8e7ce2e953d0eead4e37b6cb08fe2df64dae3d013237676d9b8ff9dbb40" Mar 21 08:44:35 crc kubenswrapper[4696]: I0321 08:44:35.338520 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerStarted","Data":"37dc28f551f9db6ffb052a7a69754f617535fa9f02cee7492942ac3ed542a742"} Mar 21 08:44:35 crc kubenswrapper[4696]: I0321 08:44:35.378247 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-64dc66d669-8685h" podStartSLOduration=8.434330851 podStartE2EDuration="24.37822315s" podCreationTimestamp="2026-03-21 08:44:11 +0000 UTC" firstStartedPulling="2026-03-21 08:44:13.258530572 +0000 UTC m=+987.379411285" lastFinishedPulling="2026-03-21 08:44:29.202422871 +0000 UTC m=+1003.323303584" observedRunningTime="2026-03-21 08:44:30.49959148 +0000 UTC m=+1004.620472193" watchObservedRunningTime="2026-03-21 08:44:35.37822315 +0000 UTC m=+1009.499103873" Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.382135 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-69d78c9949-5whvx" event={"ID":"0ecf9ee3-d5c9-4677-b264-0ddae28df715","Type":"ContainerStarted","Data":"c960b72e80015da978209db3bdbe76442553af4e2dd3b88b250a1f05e71b4ce1"} Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.383087 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-69d78c9949-5whvx" Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.384131 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-74d6f7b5c-5x9b6" event={"ID":"f593acbc-0a73-41ed-a949-02a3556dfc70","Type":"ContainerStarted","Data":"3221a036d3695fe387ec54347fabbee07556dd7ee265dc46a86db0f62eaf289a"} Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.384334 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-74d6f7b5c-5x9b6" Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.385480 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-bc5c78db9-c9skl" event={"ID":"50f56ad0-1409-4619-9b7a-887ba6997874","Type":"ContainerStarted","Data":"3cf0d342a8a5bb640b004dc2f56e8e17023b75456d02eb7f5ba92922f3c8d8b6"} Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.385649 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-bc5c78db9-c9skl" Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.387675 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-659fb58c6b-m5bx9" event={"ID":"d2672ea4-e1e0-4809-b4e9-04e2cece2179","Type":"ContainerStarted","Data":"3da55692f74b46e86c3bc58ea9d726c608e60c391b860622bd7f54379bff7854"} Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.387870 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-659fb58c6b-m5bx9" Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.389962 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvljx" event={"ID":"1ff5cfb7-37ad-4f6a-84f1-e8dbdef06084","Type":"ContainerStarted","Data":"a17b295b795eabdab0d6dc0bcb7f64d98ce4f29aeb16fa9c49af9828b92e939e"} Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.392260 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-56f74467c6-fzv5q" event={"ID":"828a9554-fa0a-4458-9d63-b9e6bd6bde49","Type":"ContainerStarted","Data":"a7095ff97b212cf833ff3519131ea2d45939db65b02d4769b0b20cc535e2483d"} Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.392453 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-56f74467c6-fzv5q" Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.393949 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" event={"ID":"026783e3-cff7-40bf-9f26-2d5e8b6006e7","Type":"ContainerStarted","Data":"c4b9e909d8885de998327600ca1131e10c13c82a1d47dab3e1afa659c786a254"} Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.394148 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.395731 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5595c7d6ff-7cg24" event={"ID":"0048e137-52d0-476b-845a-4c42829de1e2","Type":"ContainerStarted","Data":"f6436097496c04ccdd9a9e61454e8136a530d6b5f62229a6883b4854dd8bdba5"} Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.395992 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5595c7d6ff-7cg24" Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.428837 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-69d78c9949-5whvx" podStartSLOduration=3.330629473 podStartE2EDuration="30.428805041s" podCreationTimestamp="2026-03-21 08:44:11 +0000 UTC" firstStartedPulling="2026-03-21 08:44:13.556868092 +0000 UTC m=+987.677748805" lastFinishedPulling="2026-03-21 08:44:40.65504366 +0000 UTC m=+1014.775924373" observedRunningTime="2026-03-21 08:44:41.423638569 +0000 UTC m=+1015.544519292" watchObservedRunningTime="2026-03-21 08:44:41.428805041 +0000 UTC m=+1015.549685754" Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.464758 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-659fb58c6b-m5bx9" podStartSLOduration=3.172564179 podStartE2EDuration="30.464743574s" podCreationTimestamp="2026-03-21 08:44:11 +0000 UTC" firstStartedPulling="2026-03-21 08:44:13.403213608 +0000 UTC m=+987.524094321" lastFinishedPulling="2026-03-21 08:44:40.695392993 +0000 UTC m=+1014.816273716" observedRunningTime="2026-03-21 08:44:41.463035677 +0000 UTC m=+1015.583916390" watchObservedRunningTime="2026-03-21 08:44:41.464743574 +0000 UTC m=+1015.585624287" Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.466614 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kvljx" podStartSLOduration=2.317970264 podStartE2EDuration="29.466607425s" podCreationTimestamp="2026-03-21 08:44:12 +0000 UTC" firstStartedPulling="2026-03-21 08:44:13.551230677 +0000 UTC m=+987.672111390" lastFinishedPulling="2026-03-21 08:44:40.699867838 +0000 UTC m=+1014.820748551" observedRunningTime="2026-03-21 08:44:41.445029389 +0000 UTC m=+1015.565910102" watchObservedRunningTime="2026-03-21 08:44:41.466607425 +0000 UTC m=+1015.587488138" Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.495767 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-bc5c78db9-c9skl" podStartSLOduration=3.188516369 podStartE2EDuration="30.49574084s" podCreationTimestamp="2026-03-21 08:44:11 +0000 UTC" firstStartedPulling="2026-03-21 08:44:13.387997818 +0000 UTC m=+987.508878521" lastFinishedPulling="2026-03-21 08:44:40.695222279 +0000 UTC m=+1014.816102992" observedRunningTime="2026-03-21 08:44:41.488838869 +0000 UTC m=+1015.609719582" watchObservedRunningTime="2026-03-21 08:44:41.49574084 +0000 UTC m=+1015.616621553" Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.530053 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-56f74467c6-fzv5q" podStartSLOduration=3.335758786 podStartE2EDuration="30.530037958s" podCreationTimestamp="2026-03-21 08:44:11 +0000 UTC" firstStartedPulling="2026-03-21 08:44:13.509802612 +0000 UTC m=+987.630683325" lastFinishedPulling="2026-03-21 08:44:40.704081784 +0000 UTC m=+1014.824962497" observedRunningTime="2026-03-21 08:44:41.524915866 +0000 UTC m=+1015.645796579" watchObservedRunningTime="2026-03-21 08:44:41.530037958 +0000 UTC m=+1015.650918661" Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.561963 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-5595c7d6ff-7cg24" podStartSLOduration=19.607465958 podStartE2EDuration="30.561946118s" podCreationTimestamp="2026-03-21 08:44:11 +0000 UTC" firstStartedPulling="2026-03-21 08:44:29.740575194 +0000 UTC m=+1003.861455907" lastFinishedPulling="2026-03-21 08:44:40.695055354 +0000 UTC m=+1014.815936067" observedRunningTime="2026-03-21 08:44:41.550135622 +0000 UTC m=+1015.671016335" watchObservedRunningTime="2026-03-21 08:44:41.561946118 +0000 UTC m=+1015.682826831" Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.580495 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-74d6f7b5c-5x9b6" podStartSLOduration=2.396154955 podStartE2EDuration="29.580479981s" podCreationTimestamp="2026-03-21 08:44:12 +0000 UTC" firstStartedPulling="2026-03-21 08:44:13.510686037 +0000 UTC m=+987.631566750" lastFinishedPulling="2026-03-21 08:44:40.695011063 +0000 UTC m=+1014.815891776" observedRunningTime="2026-03-21 08:44:41.576494321 +0000 UTC m=+1015.697375034" watchObservedRunningTime="2026-03-21 08:44:41.580479981 +0000 UTC m=+1015.701360694" Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.619876 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" podStartSLOduration=19.549770705 podStartE2EDuration="30.619860868s" podCreationTimestamp="2026-03-21 08:44:11 +0000 UTC" firstStartedPulling="2026-03-21 08:44:29.626506434 +0000 UTC m=+1003.747387147" lastFinishedPulling="2026-03-21 08:44:40.696596597 +0000 UTC m=+1014.817477310" observedRunningTime="2026-03-21 08:44:41.617115413 +0000 UTC m=+1015.737996126" watchObservedRunningTime="2026-03-21 08:44:41.619860868 +0000 UTC m=+1015.740741581" Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.799887 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-5cfd84c587-ldj9m" Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.814180 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6d77645966-qsd27" Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.904353 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-6cc65c69fc-m8ffd" Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.924595 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7d559dcdbd-zxgmc" Mar 21 08:44:41 crc kubenswrapper[4696]: I0321 08:44:41.999482 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-6744dd545c-mvqjq" Mar 21 08:44:42 crc kubenswrapper[4696]: I0321 08:44:42.007435 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-66dd9d474d-qf5kd" Mar 21 08:44:42 crc kubenswrapper[4696]: I0321 08:44:42.009525 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-64dc66d669-8685h" Mar 21 08:44:42 crc kubenswrapper[4696]: I0321 08:44:42.035899 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6b77b7676d-vhnjl" Mar 21 08:44:42 crc kubenswrapper[4696]: I0321 08:44:42.064550 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-76b87776c9-rcz2g" Mar 21 08:44:42 crc kubenswrapper[4696]: I0321 08:44:42.142610 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-fbf7bbb96-745cq" Mar 21 08:44:42 crc kubenswrapper[4696]: I0321 08:44:42.231706 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6f5b7bcd4-pzgkp" Mar 21 08:44:42 crc kubenswrapper[4696]: I0321 08:44:42.352951 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-846c4cdcb7-4s5br" Mar 21 08:44:42 crc kubenswrapper[4696]: I0321 08:44:42.616760 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-8467ccb4c8-zk58x" Mar 21 08:44:44 crc kubenswrapper[4696]: I0321 08:44:44.292775 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-webhook-certs\") pod \"openstack-operator-controller-manager-b7fdfc6b-hrhxc\" (UID: \"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0\") " pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:44 crc kubenswrapper[4696]: I0321 08:44:44.292848 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-metrics-certs\") pod \"openstack-operator-controller-manager-b7fdfc6b-hrhxc\" (UID: \"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0\") " pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:44 crc kubenswrapper[4696]: I0321 08:44:44.299344 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-webhook-certs\") pod \"openstack-operator-controller-manager-b7fdfc6b-hrhxc\" (UID: \"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0\") " pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:44 crc kubenswrapper[4696]: I0321 08:44:44.300508 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fbbc65c6-7696-4795-8b4b-7d9c7e315bf0-metrics-certs\") pod \"openstack-operator-controller-manager-b7fdfc6b-hrhxc\" (UID: \"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0\") " pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:44 crc kubenswrapper[4696]: I0321 08:44:44.553538 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:45 crc kubenswrapper[4696]: I0321 08:44:45.089633 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc"] Mar 21 08:44:45 crc kubenswrapper[4696]: W0321 08:44:45.093925 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfbbc65c6_7696_4795_8b4b_7d9c7e315bf0.slice/crio-94a018fecf7a4001998788d009b31e6a484e407bd6c0bcc24ddb1bae03149977 WatchSource:0}: Error finding container 94a018fecf7a4001998788d009b31e6a484e407bd6c0bcc24ddb1bae03149977: Status 404 returned error can't find the container with id 94a018fecf7a4001998788d009b31e6a484e407bd6c0bcc24ddb1bae03149977 Mar 21 08:44:45 crc kubenswrapper[4696]: I0321 08:44:45.462697 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" event={"ID":"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0","Type":"ContainerStarted","Data":"69af6cb18441ca5ee9a36f6527d6749f51055cc8b1694a5eabcbc0f9036e12c7"} Mar 21 08:44:45 crc kubenswrapper[4696]: I0321 08:44:45.462975 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" event={"ID":"fbbc65c6-7696-4795-8b4b-7d9c7e315bf0","Type":"ContainerStarted","Data":"94a018fecf7a4001998788d009b31e6a484e407bd6c0bcc24ddb1bae03149977"} Mar 21 08:44:45 crc kubenswrapper[4696]: I0321 08:44:45.462990 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:44:45 crc kubenswrapper[4696]: I0321 08:44:45.489840 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" podStartSLOduration=33.489797769 podStartE2EDuration="33.489797769s" podCreationTimestamp="2026-03-21 08:44:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:44:45.488163134 +0000 UTC m=+1019.609043847" watchObservedRunningTime="2026-03-21 08:44:45.489797769 +0000 UTC m=+1019.610678482" Mar 21 08:44:46 crc kubenswrapper[4696]: I0321 08:44:46.474722 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-867f54bc44-n94f2" event={"ID":"5fe44a3f-65e2-481b-9d1d-f3b3c24944db","Type":"ContainerStarted","Data":"a14027a938d26b6bd1ab9c32738f57cde3542e0dc04513eae3ef7e35e0709c43"} Mar 21 08:44:46 crc kubenswrapper[4696]: I0321 08:44:46.475469 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-867f54bc44-n94f2" Mar 21 08:44:46 crc kubenswrapper[4696]: I0321 08:44:46.500972 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-867f54bc44-n94f2" podStartSLOduration=2.823921389 podStartE2EDuration="35.500953838s" podCreationTimestamp="2026-03-21 08:44:11 +0000 UTC" firstStartedPulling="2026-03-21 08:44:13.37721173 +0000 UTC m=+987.498092443" lastFinishedPulling="2026-03-21 08:44:46.054244179 +0000 UTC m=+1020.175124892" observedRunningTime="2026-03-21 08:44:46.494761276 +0000 UTC m=+1020.615641999" watchObservedRunningTime="2026-03-21 08:44:46.500953838 +0000 UTC m=+1020.621834551" Mar 21 08:44:47 crc kubenswrapper[4696]: I0321 08:44:47.519364 4696 scope.go:117] "RemoveContainer" containerID="2e1df926844a18d70f9120e60df866892a5d9837c576b458c331c1a881722233" Mar 21 08:44:47 crc kubenswrapper[4696]: I0321 08:44:47.626102 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5595c7d6ff-7cg24" Mar 21 08:44:47 crc kubenswrapper[4696]: I0321 08:44:47.975950 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89" Mar 21 08:44:52 crc kubenswrapper[4696]: I0321 08:44:52.315423 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-bc5c78db9-c9skl" Mar 21 08:44:52 crc kubenswrapper[4696]: I0321 08:44:52.324596 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-56f74467c6-fzv5q" Mar 21 08:44:52 crc kubenswrapper[4696]: I0321 08:44:52.414342 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-867f54bc44-n94f2" Mar 21 08:44:52 crc kubenswrapper[4696]: I0321 08:44:52.450305 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-659fb58c6b-m5bx9" Mar 21 08:44:52 crc kubenswrapper[4696]: I0321 08:44:52.484985 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-69d78c9949-5whvx" Mar 21 08:44:52 crc kubenswrapper[4696]: I0321 08:44:52.787051 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-74d6f7b5c-5x9b6" Mar 21 08:44:54 crc kubenswrapper[4696]: I0321 08:44:54.562450 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-b7fdfc6b-hrhxc" Mar 21 08:45:00 crc kubenswrapper[4696]: I0321 08:45:00.144560 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568045-xllgs"] Mar 21 08:45:00 crc kubenswrapper[4696]: I0321 08:45:00.146234 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568045-xllgs" Mar 21 08:45:00 crc kubenswrapper[4696]: I0321 08:45:00.148766 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 21 08:45:00 crc kubenswrapper[4696]: I0321 08:45:00.154200 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 21 08:45:00 crc kubenswrapper[4696]: I0321 08:45:00.166062 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568045-xllgs"] Mar 21 08:45:00 crc kubenswrapper[4696]: I0321 08:45:00.323035 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qldl9\" (UniqueName: \"kubernetes.io/projected/896b4905-5c3e-4fff-8c28-3891ae88049d-kube-api-access-qldl9\") pod \"collect-profiles-29568045-xllgs\" (UID: \"896b4905-5c3e-4fff-8c28-3891ae88049d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568045-xllgs" Mar 21 08:45:00 crc kubenswrapper[4696]: I0321 08:45:00.323085 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/896b4905-5c3e-4fff-8c28-3891ae88049d-config-volume\") pod \"collect-profiles-29568045-xllgs\" (UID: \"896b4905-5c3e-4fff-8c28-3891ae88049d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568045-xllgs" Mar 21 08:45:00 crc kubenswrapper[4696]: I0321 08:45:00.323173 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/896b4905-5c3e-4fff-8c28-3891ae88049d-secret-volume\") pod \"collect-profiles-29568045-xllgs\" (UID: \"896b4905-5c3e-4fff-8c28-3891ae88049d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568045-xllgs" Mar 21 08:45:00 crc kubenswrapper[4696]: I0321 08:45:00.425146 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/896b4905-5c3e-4fff-8c28-3891ae88049d-secret-volume\") pod \"collect-profiles-29568045-xllgs\" (UID: \"896b4905-5c3e-4fff-8c28-3891ae88049d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568045-xllgs" Mar 21 08:45:00 crc kubenswrapper[4696]: I0321 08:45:00.425235 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qldl9\" (UniqueName: \"kubernetes.io/projected/896b4905-5c3e-4fff-8c28-3891ae88049d-kube-api-access-qldl9\") pod \"collect-profiles-29568045-xllgs\" (UID: \"896b4905-5c3e-4fff-8c28-3891ae88049d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568045-xllgs" Mar 21 08:45:00 crc kubenswrapper[4696]: I0321 08:45:00.425289 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/896b4905-5c3e-4fff-8c28-3891ae88049d-config-volume\") pod \"collect-profiles-29568045-xllgs\" (UID: \"896b4905-5c3e-4fff-8c28-3891ae88049d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568045-xllgs" Mar 21 08:45:00 crc kubenswrapper[4696]: I0321 08:45:00.427422 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/896b4905-5c3e-4fff-8c28-3891ae88049d-config-volume\") pod \"collect-profiles-29568045-xllgs\" (UID: \"896b4905-5c3e-4fff-8c28-3891ae88049d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568045-xllgs" Mar 21 08:45:00 crc kubenswrapper[4696]: I0321 08:45:00.436005 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/896b4905-5c3e-4fff-8c28-3891ae88049d-secret-volume\") pod \"collect-profiles-29568045-xllgs\" (UID: \"896b4905-5c3e-4fff-8c28-3891ae88049d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568045-xllgs" Mar 21 08:45:00 crc kubenswrapper[4696]: I0321 08:45:00.446511 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qldl9\" (UniqueName: \"kubernetes.io/projected/896b4905-5c3e-4fff-8c28-3891ae88049d-kube-api-access-qldl9\") pod \"collect-profiles-29568045-xllgs\" (UID: \"896b4905-5c3e-4fff-8c28-3891ae88049d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568045-xllgs" Mar 21 08:45:00 crc kubenswrapper[4696]: I0321 08:45:00.474375 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568045-xllgs" Mar 21 08:45:00 crc kubenswrapper[4696]: I0321 08:45:00.909777 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568045-xllgs"] Mar 21 08:45:01 crc kubenswrapper[4696]: I0321 08:45:01.625342 4696 generic.go:334] "Generic (PLEG): container finished" podID="896b4905-5c3e-4fff-8c28-3891ae88049d" containerID="98aaa70452c44b76bf8ebe73361168b72c19f6b5beed5170a7ecbb55a3ee98d8" exitCode=0 Mar 21 08:45:01 crc kubenswrapper[4696]: I0321 08:45:01.625411 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568045-xllgs" event={"ID":"896b4905-5c3e-4fff-8c28-3891ae88049d","Type":"ContainerDied","Data":"98aaa70452c44b76bf8ebe73361168b72c19f6b5beed5170a7ecbb55a3ee98d8"} Mar 21 08:45:01 crc kubenswrapper[4696]: I0321 08:45:01.625439 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568045-xllgs" event={"ID":"896b4905-5c3e-4fff-8c28-3891ae88049d","Type":"ContainerStarted","Data":"7cc004ef9cda2ead2c9d7e7b9b0e06d5e6a9bca6ee3ec47cf8dd307debd2f6f3"} Mar 21 08:45:02 crc kubenswrapper[4696]: I0321 08:45:02.889809 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568045-xllgs" Mar 21 08:45:03 crc kubenswrapper[4696]: I0321 08:45:03.058045 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qldl9\" (UniqueName: \"kubernetes.io/projected/896b4905-5c3e-4fff-8c28-3891ae88049d-kube-api-access-qldl9\") pod \"896b4905-5c3e-4fff-8c28-3891ae88049d\" (UID: \"896b4905-5c3e-4fff-8c28-3891ae88049d\") " Mar 21 08:45:03 crc kubenswrapper[4696]: I0321 08:45:03.058177 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/896b4905-5c3e-4fff-8c28-3891ae88049d-secret-volume\") pod \"896b4905-5c3e-4fff-8c28-3891ae88049d\" (UID: \"896b4905-5c3e-4fff-8c28-3891ae88049d\") " Mar 21 08:45:03 crc kubenswrapper[4696]: I0321 08:45:03.058210 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/896b4905-5c3e-4fff-8c28-3891ae88049d-config-volume\") pod \"896b4905-5c3e-4fff-8c28-3891ae88049d\" (UID: \"896b4905-5c3e-4fff-8c28-3891ae88049d\") " Mar 21 08:45:03 crc kubenswrapper[4696]: I0321 08:45:03.058967 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/896b4905-5c3e-4fff-8c28-3891ae88049d-config-volume" (OuterVolumeSpecName: "config-volume") pod "896b4905-5c3e-4fff-8c28-3891ae88049d" (UID: "896b4905-5c3e-4fff-8c28-3891ae88049d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:45:03 crc kubenswrapper[4696]: I0321 08:45:03.064074 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/896b4905-5c3e-4fff-8c28-3891ae88049d-kube-api-access-qldl9" (OuterVolumeSpecName: "kube-api-access-qldl9") pod "896b4905-5c3e-4fff-8c28-3891ae88049d" (UID: "896b4905-5c3e-4fff-8c28-3891ae88049d"). InnerVolumeSpecName "kube-api-access-qldl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:45:03 crc kubenswrapper[4696]: I0321 08:45:03.070777 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/896b4905-5c3e-4fff-8c28-3891ae88049d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "896b4905-5c3e-4fff-8c28-3891ae88049d" (UID: "896b4905-5c3e-4fff-8c28-3891ae88049d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:45:03 crc kubenswrapper[4696]: I0321 08:45:03.160102 4696 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/896b4905-5c3e-4fff-8c28-3891ae88049d-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 21 08:45:03 crc kubenswrapper[4696]: I0321 08:45:03.160135 4696 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/896b4905-5c3e-4fff-8c28-3891ae88049d-config-volume\") on node \"crc\" DevicePath \"\"" Mar 21 08:45:03 crc kubenswrapper[4696]: I0321 08:45:03.160263 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qldl9\" (UniqueName: \"kubernetes.io/projected/896b4905-5c3e-4fff-8c28-3891ae88049d-kube-api-access-qldl9\") on node \"crc\" DevicePath \"\"" Mar 21 08:45:03 crc kubenswrapper[4696]: I0321 08:45:03.644064 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568045-xllgs" event={"ID":"896b4905-5c3e-4fff-8c28-3891ae88049d","Type":"ContainerDied","Data":"7cc004ef9cda2ead2c9d7e7b9b0e06d5e6a9bca6ee3ec47cf8dd307debd2f6f3"} Mar 21 08:45:03 crc kubenswrapper[4696]: I0321 08:45:03.644108 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7cc004ef9cda2ead2c9d7e7b9b0e06d5e6a9bca6ee3ec47cf8dd307debd2f6f3" Mar 21 08:45:03 crc kubenswrapper[4696]: I0321 08:45:03.644130 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568045-xllgs" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.683691 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4xhpm"] Mar 21 08:45:11 crc kubenswrapper[4696]: E0321 08:45:11.684603 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="896b4905-5c3e-4fff-8c28-3891ae88049d" containerName="collect-profiles" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.684619 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="896b4905-5c3e-4fff-8c28-3891ae88049d" containerName="collect-profiles" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.684787 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="896b4905-5c3e-4fff-8c28-3891ae88049d" containerName="collect-profiles" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.685660 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-4xhpm" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.689390 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-g74vm" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.689712 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.689894 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.690071 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.698472 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4xhpm"] Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.741158 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-hxprq"] Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.751659 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-hxprq" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.757024 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.767169 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-hxprq"] Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.886322 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26009008-54e8-48fc-b1b5-99c0e0c1e544-config\") pod \"dnsmasq-dns-78dd6ddcc-hxprq\" (UID: \"26009008-54e8-48fc-b1b5-99c0e0c1e544\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hxprq" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.886365 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrk4w\" (UniqueName: \"kubernetes.io/projected/26009008-54e8-48fc-b1b5-99c0e0c1e544-kube-api-access-nrk4w\") pod \"dnsmasq-dns-78dd6ddcc-hxprq\" (UID: \"26009008-54e8-48fc-b1b5-99c0e0c1e544\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hxprq" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.886403 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31577bd-23ed-4ab0-8cf1-42a40b0cc941-config\") pod \"dnsmasq-dns-675f4bcbfc-4xhpm\" (UID: \"c31577bd-23ed-4ab0-8cf1-42a40b0cc941\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4xhpm" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.886456 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26009008-54e8-48fc-b1b5-99c0e0c1e544-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-hxprq\" (UID: \"26009008-54e8-48fc-b1b5-99c0e0c1e544\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hxprq" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.886628 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccz2l\" (UniqueName: \"kubernetes.io/projected/c31577bd-23ed-4ab0-8cf1-42a40b0cc941-kube-api-access-ccz2l\") pod \"dnsmasq-dns-675f4bcbfc-4xhpm\" (UID: \"c31577bd-23ed-4ab0-8cf1-42a40b0cc941\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4xhpm" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.992097 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26009008-54e8-48fc-b1b5-99c0e0c1e544-config\") pod \"dnsmasq-dns-78dd6ddcc-hxprq\" (UID: \"26009008-54e8-48fc-b1b5-99c0e0c1e544\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hxprq" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.992180 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrk4w\" (UniqueName: \"kubernetes.io/projected/26009008-54e8-48fc-b1b5-99c0e0c1e544-kube-api-access-nrk4w\") pod \"dnsmasq-dns-78dd6ddcc-hxprq\" (UID: \"26009008-54e8-48fc-b1b5-99c0e0c1e544\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hxprq" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.992241 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31577bd-23ed-4ab0-8cf1-42a40b0cc941-config\") pod \"dnsmasq-dns-675f4bcbfc-4xhpm\" (UID: \"c31577bd-23ed-4ab0-8cf1-42a40b0cc941\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4xhpm" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.992275 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26009008-54e8-48fc-b1b5-99c0e0c1e544-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-hxprq\" (UID: \"26009008-54e8-48fc-b1b5-99c0e0c1e544\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hxprq" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.992343 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccz2l\" (UniqueName: \"kubernetes.io/projected/c31577bd-23ed-4ab0-8cf1-42a40b0cc941-kube-api-access-ccz2l\") pod \"dnsmasq-dns-675f4bcbfc-4xhpm\" (UID: \"c31577bd-23ed-4ab0-8cf1-42a40b0cc941\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4xhpm" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.993170 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26009008-54e8-48fc-b1b5-99c0e0c1e544-config\") pod \"dnsmasq-dns-78dd6ddcc-hxprq\" (UID: \"26009008-54e8-48fc-b1b5-99c0e0c1e544\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hxprq" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.994001 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26009008-54e8-48fc-b1b5-99c0e0c1e544-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-hxprq\" (UID: \"26009008-54e8-48fc-b1b5-99c0e0c1e544\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hxprq" Mar 21 08:45:11 crc kubenswrapper[4696]: I0321 08:45:11.994084 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31577bd-23ed-4ab0-8cf1-42a40b0cc941-config\") pod \"dnsmasq-dns-675f4bcbfc-4xhpm\" (UID: \"c31577bd-23ed-4ab0-8cf1-42a40b0cc941\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4xhpm" Mar 21 08:45:12 crc kubenswrapper[4696]: I0321 08:45:12.014222 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccz2l\" (UniqueName: \"kubernetes.io/projected/c31577bd-23ed-4ab0-8cf1-42a40b0cc941-kube-api-access-ccz2l\") pod \"dnsmasq-dns-675f4bcbfc-4xhpm\" (UID: \"c31577bd-23ed-4ab0-8cf1-42a40b0cc941\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4xhpm" Mar 21 08:45:12 crc kubenswrapper[4696]: I0321 08:45:12.019130 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrk4w\" (UniqueName: \"kubernetes.io/projected/26009008-54e8-48fc-b1b5-99c0e0c1e544-kube-api-access-nrk4w\") pod \"dnsmasq-dns-78dd6ddcc-hxprq\" (UID: \"26009008-54e8-48fc-b1b5-99c0e0c1e544\") " pod="openstack/dnsmasq-dns-78dd6ddcc-hxprq" Mar 21 08:45:12 crc kubenswrapper[4696]: I0321 08:45:12.081910 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-hxprq" Mar 21 08:45:12 crc kubenswrapper[4696]: I0321 08:45:12.309617 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-4xhpm" Mar 21 08:45:12 crc kubenswrapper[4696]: I0321 08:45:12.487575 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-hxprq"] Mar 21 08:45:12 crc kubenswrapper[4696]: W0321 08:45:12.490420 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26009008_54e8_48fc_b1b5_99c0e0c1e544.slice/crio-d19214f1fe3e7145632c88ea346e1722182855d39c172586721d90f457105b56 WatchSource:0}: Error finding container d19214f1fe3e7145632c88ea346e1722182855d39c172586721d90f457105b56: Status 404 returned error can't find the container with id d19214f1fe3e7145632c88ea346e1722182855d39c172586721d90f457105b56 Mar 21 08:45:12 crc kubenswrapper[4696]: W0321 08:45:12.709428 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc31577bd_23ed_4ab0_8cf1_42a40b0cc941.slice/crio-9084f0474616d92968667f58dd6404e15d4363745f600cb2f5e1805478d5a25e WatchSource:0}: Error finding container 9084f0474616d92968667f58dd6404e15d4363745f600cb2f5e1805478d5a25e: Status 404 returned error can't find the container with id 9084f0474616d92968667f58dd6404e15d4363745f600cb2f5e1805478d5a25e Mar 21 08:45:12 crc kubenswrapper[4696]: I0321 08:45:12.709656 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4xhpm"] Mar 21 08:45:12 crc kubenswrapper[4696]: I0321 08:45:12.736433 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-4xhpm" event={"ID":"c31577bd-23ed-4ab0-8cf1-42a40b0cc941","Type":"ContainerStarted","Data":"9084f0474616d92968667f58dd6404e15d4363745f600cb2f5e1805478d5a25e"} Mar 21 08:45:12 crc kubenswrapper[4696]: I0321 08:45:12.738041 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-hxprq" event={"ID":"26009008-54e8-48fc-b1b5-99c0e0c1e544","Type":"ContainerStarted","Data":"d19214f1fe3e7145632c88ea346e1722182855d39c172586721d90f457105b56"} Mar 21 08:45:14 crc kubenswrapper[4696]: I0321 08:45:14.594830 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4xhpm"] Mar 21 08:45:14 crc kubenswrapper[4696]: I0321 08:45:14.610616 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-27j2f"] Mar 21 08:45:14 crc kubenswrapper[4696]: I0321 08:45:14.612162 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" Mar 21 08:45:14 crc kubenswrapper[4696]: I0321 08:45:14.622008 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-27j2f"] Mar 21 08:45:14 crc kubenswrapper[4696]: I0321 08:45:14.655980 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnr7n\" (UniqueName: \"kubernetes.io/projected/065e4107-c490-498d-a0f0-fce4011c38f1-kube-api-access-tnr7n\") pod \"dnsmasq-dns-5ccc8479f9-27j2f\" (UID: \"065e4107-c490-498d-a0f0-fce4011c38f1\") " pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" Mar 21 08:45:14 crc kubenswrapper[4696]: I0321 08:45:14.656118 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/065e4107-c490-498d-a0f0-fce4011c38f1-config\") pod \"dnsmasq-dns-5ccc8479f9-27j2f\" (UID: \"065e4107-c490-498d-a0f0-fce4011c38f1\") " pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" Mar 21 08:45:14 crc kubenswrapper[4696]: I0321 08:45:14.656233 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/065e4107-c490-498d-a0f0-fce4011c38f1-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-27j2f\" (UID: \"065e4107-c490-498d-a0f0-fce4011c38f1\") " pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" Mar 21 08:45:14 crc kubenswrapper[4696]: I0321 08:45:14.758410 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/065e4107-c490-498d-a0f0-fce4011c38f1-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-27j2f\" (UID: \"065e4107-c490-498d-a0f0-fce4011c38f1\") " pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" Mar 21 08:45:14 crc kubenswrapper[4696]: I0321 08:45:14.758474 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnr7n\" (UniqueName: \"kubernetes.io/projected/065e4107-c490-498d-a0f0-fce4011c38f1-kube-api-access-tnr7n\") pod \"dnsmasq-dns-5ccc8479f9-27j2f\" (UID: \"065e4107-c490-498d-a0f0-fce4011c38f1\") " pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" Mar 21 08:45:14 crc kubenswrapper[4696]: I0321 08:45:14.758548 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/065e4107-c490-498d-a0f0-fce4011c38f1-config\") pod \"dnsmasq-dns-5ccc8479f9-27j2f\" (UID: \"065e4107-c490-498d-a0f0-fce4011c38f1\") " pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" Mar 21 08:45:14 crc kubenswrapper[4696]: I0321 08:45:14.759597 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/065e4107-c490-498d-a0f0-fce4011c38f1-config\") pod \"dnsmasq-dns-5ccc8479f9-27j2f\" (UID: \"065e4107-c490-498d-a0f0-fce4011c38f1\") " pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" Mar 21 08:45:14 crc kubenswrapper[4696]: I0321 08:45:14.760267 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/065e4107-c490-498d-a0f0-fce4011c38f1-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-27j2f\" (UID: \"065e4107-c490-498d-a0f0-fce4011c38f1\") " pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" Mar 21 08:45:14 crc kubenswrapper[4696]: I0321 08:45:14.780320 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnr7n\" (UniqueName: \"kubernetes.io/projected/065e4107-c490-498d-a0f0-fce4011c38f1-kube-api-access-tnr7n\") pod \"dnsmasq-dns-5ccc8479f9-27j2f\" (UID: \"065e4107-c490-498d-a0f0-fce4011c38f1\") " pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" Mar 21 08:45:14 crc kubenswrapper[4696]: I0321 08:45:14.875035 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-hxprq"] Mar 21 08:45:14 crc kubenswrapper[4696]: I0321 08:45:14.896358 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bvmvs"] Mar 21 08:45:14 crc kubenswrapper[4696]: I0321 08:45:14.902495 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" Mar 21 08:45:14 crc kubenswrapper[4696]: I0321 08:45:14.913156 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bvmvs"] Mar 21 08:45:14 crc kubenswrapper[4696]: I0321 08:45:14.932751 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.067272 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-bvmvs\" (UID: \"4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc\") " pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.067311 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc-config\") pod \"dnsmasq-dns-57d769cc4f-bvmvs\" (UID: \"4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc\") " pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.067515 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-882nj\" (UniqueName: \"kubernetes.io/projected/4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc-kube-api-access-882nj\") pod \"dnsmasq-dns-57d769cc4f-bvmvs\" (UID: \"4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc\") " pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.169541 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-bvmvs\" (UID: \"4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc\") " pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.169595 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc-config\") pod \"dnsmasq-dns-57d769cc4f-bvmvs\" (UID: \"4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc\") " pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.169652 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-882nj\" (UniqueName: \"kubernetes.io/projected/4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc-kube-api-access-882nj\") pod \"dnsmasq-dns-57d769cc4f-bvmvs\" (UID: \"4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc\") " pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.170916 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-bvmvs\" (UID: \"4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc\") " pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.172315 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc-config\") pod \"dnsmasq-dns-57d769cc4f-bvmvs\" (UID: \"4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc\") " pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.187606 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-882nj\" (UniqueName: \"kubernetes.io/projected/4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc-kube-api-access-882nj\") pod \"dnsmasq-dns-57d769cc4f-bvmvs\" (UID: \"4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc\") " pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.226451 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.381695 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-27j2f"] Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.757226 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.758420 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.761589 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.761678 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.761589 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-mxjmk" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.761589 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.761902 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.762015 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.762179 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.770853 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.879781 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.879880 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b057ad21-8030-49e9-b400-c36d433b9f8c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.879904 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmw6t\" (UniqueName: \"kubernetes.io/projected/b057ad21-8030-49e9-b400-c36d433b9f8c-kube-api-access-dmw6t\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.879923 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f47b560c-4e99-44f7-928a-39b1f9203a25\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f47b560c-4e99-44f7-928a-39b1f9203a25\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.879943 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.879966 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b057ad21-8030-49e9-b400-c36d433b9f8c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.879986 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b057ad21-8030-49e9-b400-c36d433b9f8c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.880024 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b057ad21-8030-49e9-b400-c36d433b9f8c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.880058 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.880081 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.880104 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b057ad21-8030-49e9-b400-c36d433b9f8c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.981295 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b057ad21-8030-49e9-b400-c36d433b9f8c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.981357 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmw6t\" (UniqueName: \"kubernetes.io/projected/b057ad21-8030-49e9-b400-c36d433b9f8c-kube-api-access-dmw6t\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.981388 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f47b560c-4e99-44f7-928a-39b1f9203a25\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f47b560c-4e99-44f7-928a-39b1f9203a25\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.981410 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.981440 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b057ad21-8030-49e9-b400-c36d433b9f8c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.981468 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b057ad21-8030-49e9-b400-c36d433b9f8c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.981523 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b057ad21-8030-49e9-b400-c36d433b9f8c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.981564 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.981598 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.981633 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b057ad21-8030-49e9-b400-c36d433b9f8c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.981664 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.982331 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.982467 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b057ad21-8030-49e9-b400-c36d433b9f8c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.982633 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.983001 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b057ad21-8030-49e9-b400-c36d433b9f8c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.983385 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b057ad21-8030-49e9-b400-c36d433b9f8c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.984459 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.984501 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f47b560c-4e99-44f7-928a-39b1f9203a25\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f47b560c-4e99-44f7-928a-39b1f9203a25\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/3d548c5607941a22cfe7a5753cb098dc630cefe69835678ddcd11d0cea80b16b/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.986287 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.986287 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b057ad21-8030-49e9-b400-c36d433b9f8c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:15 crc kubenswrapper[4696]: I0321 08:45:15.990248 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.001315 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b057ad21-8030-49e9-b400-c36d433b9f8c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.002756 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmw6t\" (UniqueName: \"kubernetes.io/projected/b057ad21-8030-49e9-b400-c36d433b9f8c-kube-api-access-dmw6t\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.021671 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f47b560c-4e99-44f7-928a-39b1f9203a25\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f47b560c-4e99-44f7-928a-39b1f9203a25\") pod \"rabbitmq-cell1-server-0\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.054459 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.056365 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.062279 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.062719 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-hzkmd" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.062898 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.063187 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.063359 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.063593 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.065411 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.069919 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.086851 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.185498 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.185553 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.185625 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.185645 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.185689 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.185715 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.185735 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.185754 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-config-data\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.185769 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.185793 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.185810 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjtkr\" (UniqueName: \"kubernetes.io/projected/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-kube-api-access-gjtkr\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.287457 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.287507 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-config-data\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.287526 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.287550 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.287567 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjtkr\" (UniqueName: \"kubernetes.io/projected/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-kube-api-access-gjtkr\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.287608 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.287638 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.287672 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.287687 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.287752 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.287806 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.288255 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.288662 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.289440 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.289838 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-config-data\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.291084 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.293485 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.293519 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/087398aff111d894aab0c73f079ac83ad54622acef4ed647806a18bd4dd4491c/globalmount\"" pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.301469 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.302316 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.303262 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.304800 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.306368 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjtkr\" (UniqueName: \"kubernetes.io/projected/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-kube-api-access-gjtkr\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.330246 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d\") pod \"rabbitmq-server-0\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " pod="openstack/rabbitmq-server-0" Mar 21 08:45:16 crc kubenswrapper[4696]: I0321 08:45:16.385417 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.272375 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.274468 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.276667 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.278811 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.279269 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-nlcch" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.279610 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.284881 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.293916 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.427501 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/919a11c7-cf69-4660-8de6-6b1a54b63f26-kolla-config\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.427588 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-39507957-8d2a-482e-b339-fed68c41419d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-39507957-8d2a-482e-b339-fed68c41419d\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.427611 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/919a11c7-cf69-4660-8de6-6b1a54b63f26-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.427637 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919a11c7-cf69-4660-8de6-6b1a54b63f26-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.427678 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/919a11c7-cf69-4660-8de6-6b1a54b63f26-config-data-default\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.427694 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6gkf\" (UniqueName: \"kubernetes.io/projected/919a11c7-cf69-4660-8de6-6b1a54b63f26-kube-api-access-v6gkf\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.427931 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/919a11c7-cf69-4660-8de6-6b1a54b63f26-config-data-generated\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.427989 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/919a11c7-cf69-4660-8de6-6b1a54b63f26-operator-scripts\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.529641 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/919a11c7-cf69-4660-8de6-6b1a54b63f26-operator-scripts\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.529708 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/919a11c7-cf69-4660-8de6-6b1a54b63f26-kolla-config\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.529775 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/919a11c7-cf69-4660-8de6-6b1a54b63f26-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.531142 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-39507957-8d2a-482e-b339-fed68c41419d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-39507957-8d2a-482e-b339-fed68c41419d\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.532387 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919a11c7-cf69-4660-8de6-6b1a54b63f26-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.532540 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/919a11c7-cf69-4660-8de6-6b1a54b63f26-config-data-default\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.532853 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6gkf\" (UniqueName: \"kubernetes.io/projected/919a11c7-cf69-4660-8de6-6b1a54b63f26-kube-api-access-v6gkf\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.532965 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/919a11c7-cf69-4660-8de6-6b1a54b63f26-config-data-generated\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.532984 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/919a11c7-cf69-4660-8de6-6b1a54b63f26-operator-scripts\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.534158 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/919a11c7-cf69-4660-8de6-6b1a54b63f26-config-data-generated\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.534807 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/919a11c7-cf69-4660-8de6-6b1a54b63f26-config-data-default\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.535335 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/919a11c7-cf69-4660-8de6-6b1a54b63f26-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.546108 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/919a11c7-cf69-4660-8de6-6b1a54b63f26-kolla-config\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.553254 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.553294 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-39507957-8d2a-482e-b339-fed68c41419d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-39507957-8d2a-482e-b339-fed68c41419d\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e2618b987a7483796001e87b1d548e17df7a4a909e2fc3b87e4b79c1899af33f/globalmount\"" pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.554018 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6gkf\" (UniqueName: \"kubernetes.io/projected/919a11c7-cf69-4660-8de6-6b1a54b63f26-kube-api-access-v6gkf\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.554501 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919a11c7-cf69-4660-8de6-6b1a54b63f26-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.580035 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-39507957-8d2a-482e-b339-fed68c41419d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-39507957-8d2a-482e-b339-fed68c41419d\") pod \"openstack-galera-0\" (UID: \"919a11c7-cf69-4660-8de6-6b1a54b63f26\") " pod="openstack/openstack-galera-0" Mar 21 08:45:17 crc kubenswrapper[4696]: I0321 08:45:17.597199 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.483381 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.484641 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.486689 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.486857 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-bb4tq" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.488010 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.491478 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.500556 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.650730 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-93b74efc-4f0e-414d-821c-09c8063247d7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93b74efc-4f0e-414d-821c-09c8063247d7\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.650786 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57e750dc-d19c-45d4-9905-b0a15a6a642d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.650807 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/57e750dc-d19c-45d4-9905-b0a15a6a642d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.650893 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57e750dc-d19c-45d4-9905-b0a15a6a642d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.650911 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/57e750dc-d19c-45d4-9905-b0a15a6a642d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.650947 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/57e750dc-d19c-45d4-9905-b0a15a6a642d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.650966 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/57e750dc-d19c-45d4-9905-b0a15a6a642d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.650998 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtmcd\" (UniqueName: \"kubernetes.io/projected/57e750dc-d19c-45d4-9905-b0a15a6a642d-kube-api-access-vtmcd\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.693934 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.695207 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.699774 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-xchdg" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.699981 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.702222 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.707161 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.752639 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-93b74efc-4f0e-414d-821c-09c8063247d7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93b74efc-4f0e-414d-821c-09c8063247d7\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.752699 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57e750dc-d19c-45d4-9905-b0a15a6a642d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.752727 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/57e750dc-d19c-45d4-9905-b0a15a6a642d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.752794 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57e750dc-d19c-45d4-9905-b0a15a6a642d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.752830 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/57e750dc-d19c-45d4-9905-b0a15a6a642d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.752866 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/57e750dc-d19c-45d4-9905-b0a15a6a642d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.752889 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/57e750dc-d19c-45d4-9905-b0a15a6a642d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.752920 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtmcd\" (UniqueName: \"kubernetes.io/projected/57e750dc-d19c-45d4-9905-b0a15a6a642d-kube-api-access-vtmcd\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.753500 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/57e750dc-d19c-45d4-9905-b0a15a6a642d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.753555 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/57e750dc-d19c-45d4-9905-b0a15a6a642d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.753785 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/57e750dc-d19c-45d4-9905-b0a15a6a642d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.754641 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.754679 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-93b74efc-4f0e-414d-821c-09c8063247d7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93b74efc-4f0e-414d-821c-09c8063247d7\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1dc550f0c86070baace3d931efabe864db934be52ee668e717b6c37ccca7a361/globalmount\"" pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.755252 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57e750dc-d19c-45d4-9905-b0a15a6a642d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.766703 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/57e750dc-d19c-45d4-9905-b0a15a6a642d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.769130 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtmcd\" (UniqueName: \"kubernetes.io/projected/57e750dc-d19c-45d4-9905-b0a15a6a642d-kube-api-access-vtmcd\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.774969 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57e750dc-d19c-45d4-9905-b0a15a6a642d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.789109 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-93b74efc-4f0e-414d-821c-09c8063247d7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93b74efc-4f0e-414d-821c-09c8063247d7\") pod \"openstack-cell1-galera-0\" (UID: \"57e750dc-d19c-45d4-9905-b0a15a6a642d\") " pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.791715 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" event={"ID":"065e4107-c490-498d-a0f0-fce4011c38f1","Type":"ContainerStarted","Data":"32e469f0a3f6c7870fde7ab3d4bd0355051fe89b9f5755da50661bd7b0af38fd"} Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.817110 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.854220 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2dzh\" (UniqueName: \"kubernetes.io/projected/dd0d5d6d-801d-4446-a325-f82b2d55bca6-kube-api-access-z2dzh\") pod \"memcached-0\" (UID: \"dd0d5d6d-801d-4446-a325-f82b2d55bca6\") " pod="openstack/memcached-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.854266 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd0d5d6d-801d-4446-a325-f82b2d55bca6-memcached-tls-certs\") pod \"memcached-0\" (UID: \"dd0d5d6d-801d-4446-a325-f82b2d55bca6\") " pod="openstack/memcached-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.854304 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd0d5d6d-801d-4446-a325-f82b2d55bca6-combined-ca-bundle\") pod \"memcached-0\" (UID: \"dd0d5d6d-801d-4446-a325-f82b2d55bca6\") " pod="openstack/memcached-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.854339 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd0d5d6d-801d-4446-a325-f82b2d55bca6-config-data\") pod \"memcached-0\" (UID: \"dd0d5d6d-801d-4446-a325-f82b2d55bca6\") " pod="openstack/memcached-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.854376 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dd0d5d6d-801d-4446-a325-f82b2d55bca6-kolla-config\") pod \"memcached-0\" (UID: \"dd0d5d6d-801d-4446-a325-f82b2d55bca6\") " pod="openstack/memcached-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.955660 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2dzh\" (UniqueName: \"kubernetes.io/projected/dd0d5d6d-801d-4446-a325-f82b2d55bca6-kube-api-access-z2dzh\") pod \"memcached-0\" (UID: \"dd0d5d6d-801d-4446-a325-f82b2d55bca6\") " pod="openstack/memcached-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.955711 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd0d5d6d-801d-4446-a325-f82b2d55bca6-memcached-tls-certs\") pod \"memcached-0\" (UID: \"dd0d5d6d-801d-4446-a325-f82b2d55bca6\") " pod="openstack/memcached-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.955749 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd0d5d6d-801d-4446-a325-f82b2d55bca6-combined-ca-bundle\") pod \"memcached-0\" (UID: \"dd0d5d6d-801d-4446-a325-f82b2d55bca6\") " pod="openstack/memcached-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.955786 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd0d5d6d-801d-4446-a325-f82b2d55bca6-config-data\") pod \"memcached-0\" (UID: \"dd0d5d6d-801d-4446-a325-f82b2d55bca6\") " pod="openstack/memcached-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.955851 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dd0d5d6d-801d-4446-a325-f82b2d55bca6-kolla-config\") pod \"memcached-0\" (UID: \"dd0d5d6d-801d-4446-a325-f82b2d55bca6\") " pod="openstack/memcached-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.956640 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dd0d5d6d-801d-4446-a325-f82b2d55bca6-kolla-config\") pod \"memcached-0\" (UID: \"dd0d5d6d-801d-4446-a325-f82b2d55bca6\") " pod="openstack/memcached-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.956714 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd0d5d6d-801d-4446-a325-f82b2d55bca6-config-data\") pod \"memcached-0\" (UID: \"dd0d5d6d-801d-4446-a325-f82b2d55bca6\") " pod="openstack/memcached-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.959536 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd0d5d6d-801d-4446-a325-f82b2d55bca6-combined-ca-bundle\") pod \"memcached-0\" (UID: \"dd0d5d6d-801d-4446-a325-f82b2d55bca6\") " pod="openstack/memcached-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.964963 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd0d5d6d-801d-4446-a325-f82b2d55bca6-memcached-tls-certs\") pod \"memcached-0\" (UID: \"dd0d5d6d-801d-4446-a325-f82b2d55bca6\") " pod="openstack/memcached-0" Mar 21 08:45:18 crc kubenswrapper[4696]: I0321 08:45:18.983370 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2dzh\" (UniqueName: \"kubernetes.io/projected/dd0d5d6d-801d-4446-a325-f82b2d55bca6-kube-api-access-z2dzh\") pod \"memcached-0\" (UID: \"dd0d5d6d-801d-4446-a325-f82b2d55bca6\") " pod="openstack/memcached-0" Mar 21 08:45:19 crc kubenswrapper[4696]: I0321 08:45:19.013406 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 21 08:45:20 crc kubenswrapper[4696]: I0321 08:45:20.779320 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 21 08:45:20 crc kubenswrapper[4696]: I0321 08:45:20.780573 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 21 08:45:20 crc kubenswrapper[4696]: I0321 08:45:20.787700 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-mdcrk" Mar 21 08:45:20 crc kubenswrapper[4696]: I0321 08:45:20.794578 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 21 08:45:20 crc kubenswrapper[4696]: I0321 08:45:20.889341 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dl2g\" (UniqueName: \"kubernetes.io/projected/c1b37b82-bd6a-4e10-98cd-89fbff6bb554-kube-api-access-5dl2g\") pod \"kube-state-metrics-0\" (UID: \"c1b37b82-bd6a-4e10-98cd-89fbff6bb554\") " pod="openstack/kube-state-metrics-0" Mar 21 08:45:20 crc kubenswrapper[4696]: I0321 08:45:20.991109 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dl2g\" (UniqueName: \"kubernetes.io/projected/c1b37b82-bd6a-4e10-98cd-89fbff6bb554-kube-api-access-5dl2g\") pod \"kube-state-metrics-0\" (UID: \"c1b37b82-bd6a-4e10-98cd-89fbff6bb554\") " pod="openstack/kube-state-metrics-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.027213 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dl2g\" (UniqueName: \"kubernetes.io/projected/c1b37b82-bd6a-4e10-98cd-89fbff6bb554-kube-api-access-5dl2g\") pod \"kube-state-metrics-0\" (UID: \"c1b37b82-bd6a-4e10-98cd-89fbff6bb554\") " pod="openstack/kube-state-metrics-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.100423 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.442183 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.443979 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.445563 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.445765 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-2ldmg" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.445858 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.448057 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.449170 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.462558 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.598903 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b1808d81-bf6f-4201-972a-507d6b41f33f-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"b1808d81-bf6f-4201-972a-507d6b41f33f\") " pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.598983 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pth8v\" (UniqueName: \"kubernetes.io/projected/b1808d81-bf6f-4201-972a-507d6b41f33f-kube-api-access-pth8v\") pod \"alertmanager-metric-storage-0\" (UID: \"b1808d81-bf6f-4201-972a-507d6b41f33f\") " pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.599015 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/b1808d81-bf6f-4201-972a-507d6b41f33f-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"b1808d81-bf6f-4201-972a-507d6b41f33f\") " pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.599052 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b1808d81-bf6f-4201-972a-507d6b41f33f-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"b1808d81-bf6f-4201-972a-507d6b41f33f\") " pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.599084 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/b1808d81-bf6f-4201-972a-507d6b41f33f-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"b1808d81-bf6f-4201-972a-507d6b41f33f\") " pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.599132 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/b1808d81-bf6f-4201-972a-507d6b41f33f-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"b1808d81-bf6f-4201-972a-507d6b41f33f\") " pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.599154 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b1808d81-bf6f-4201-972a-507d6b41f33f-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"b1808d81-bf6f-4201-972a-507d6b41f33f\") " pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.700189 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/b1808d81-bf6f-4201-972a-507d6b41f33f-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"b1808d81-bf6f-4201-972a-507d6b41f33f\") " pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.700247 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b1808d81-bf6f-4201-972a-507d6b41f33f-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"b1808d81-bf6f-4201-972a-507d6b41f33f\") " pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.700306 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b1808d81-bf6f-4201-972a-507d6b41f33f-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"b1808d81-bf6f-4201-972a-507d6b41f33f\") " pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.700400 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pth8v\" (UniqueName: \"kubernetes.io/projected/b1808d81-bf6f-4201-972a-507d6b41f33f-kube-api-access-pth8v\") pod \"alertmanager-metric-storage-0\" (UID: \"b1808d81-bf6f-4201-972a-507d6b41f33f\") " pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.700438 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/b1808d81-bf6f-4201-972a-507d6b41f33f-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"b1808d81-bf6f-4201-972a-507d6b41f33f\") " pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.700533 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b1808d81-bf6f-4201-972a-507d6b41f33f-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"b1808d81-bf6f-4201-972a-507d6b41f33f\") " pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.700592 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/b1808d81-bf6f-4201-972a-507d6b41f33f-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"b1808d81-bf6f-4201-972a-507d6b41f33f\") " pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.702052 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/b1808d81-bf6f-4201-972a-507d6b41f33f-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"b1808d81-bf6f-4201-972a-507d6b41f33f\") " pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.708473 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b1808d81-bf6f-4201-972a-507d6b41f33f-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"b1808d81-bf6f-4201-972a-507d6b41f33f\") " pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.708598 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b1808d81-bf6f-4201-972a-507d6b41f33f-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"b1808d81-bf6f-4201-972a-507d6b41f33f\") " pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.721411 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/b1808d81-bf6f-4201-972a-507d6b41f33f-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"b1808d81-bf6f-4201-972a-507d6b41f33f\") " pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.721839 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/b1808d81-bf6f-4201-972a-507d6b41f33f-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"b1808d81-bf6f-4201-972a-507d6b41f33f\") " pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.733553 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b1808d81-bf6f-4201-972a-507d6b41f33f-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"b1808d81-bf6f-4201-972a-507d6b41f33f\") " pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.741362 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pth8v\" (UniqueName: \"kubernetes.io/projected/b1808d81-bf6f-4201-972a-507d6b41f33f-kube-api-access-pth8v\") pod \"alertmanager-metric-storage-0\" (UID: \"b1808d81-bf6f-4201-972a-507d6b41f33f\") " pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:21 crc kubenswrapper[4696]: I0321 08:45:21.809230 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.231243 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bvmvs"] Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.554463 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.556320 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.560326 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.560385 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-1" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.560410 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-v6ljh" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.560442 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.560401 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.560561 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.561005 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-2" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.578659 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.582675 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.717467 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.717579 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-af131abf-8240-4971-8fb5-3805d310c306\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af131abf-8240-4971-8fb5-3805d310c306\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.717625 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.717652 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lqpr\" (UniqueName: \"kubernetes.io/projected/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-kube-api-access-9lqpr\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.717681 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.717721 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.717893 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.718041 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.718132 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-config\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.718317 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.820049 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.820109 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lqpr\" (UniqueName: \"kubernetes.io/projected/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-kube-api-access-9lqpr\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.820132 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.820161 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.820181 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.820227 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.820258 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-config\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.820276 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.820324 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.820385 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-af131abf-8240-4971-8fb5-3805d310c306\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af131abf-8240-4971-8fb5-3805d310c306\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.821635 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.821870 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.822509 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.823050 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.823076 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-af131abf-8240-4971-8fb5-3805d310c306\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af131abf-8240-4971-8fb5-3805d310c306\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d67ff30a3d5f27596f87a4e17cfdebdac6e352f692a21f0c4f6bda3eb4f6d98f/globalmount\"" pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.826714 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.832850 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-config\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.834326 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.836711 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.841517 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.841675 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lqpr\" (UniqueName: \"kubernetes.io/projected/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-kube-api-access-9lqpr\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:22 crc kubenswrapper[4696]: I0321 08:45:22.895568 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-af131abf-8240-4971-8fb5-3805d310c306\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af131abf-8240-4971-8fb5-3805d310c306\") pod \"prometheus-metric-storage-0\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:23 crc kubenswrapper[4696]: I0321 08:45:23.187737 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.524788 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4vd4t"] Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.526364 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.528235 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.528883 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.529003 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-42ccb" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.541565 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-dmbpm"] Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.543299 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.560680 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4vd4t"] Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.587639 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-dmbpm"] Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.667000 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a-combined-ca-bundle\") pod \"ovn-controller-4vd4t\" (UID: \"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a\") " pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.667053 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a-scripts\") pod \"ovn-controller-4vd4t\" (UID: \"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a\") " pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.667072 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7f0d9af4-d6e3-4168-b83d-5a482fb51d44-var-lib\") pod \"ovn-controller-ovs-dmbpm\" (UID: \"7f0d9af4-d6e3-4168-b83d-5a482fb51d44\") " pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.667095 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a-ovn-controller-tls-certs\") pod \"ovn-controller-4vd4t\" (UID: \"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a\") " pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.667117 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f0d9af4-d6e3-4168-b83d-5a482fb51d44-scripts\") pod \"ovn-controller-ovs-dmbpm\" (UID: \"7f0d9af4-d6e3-4168-b83d-5a482fb51d44\") " pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.667134 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a-var-run-ovn\") pod \"ovn-controller-4vd4t\" (UID: \"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a\") " pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.667199 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7f0d9af4-d6e3-4168-b83d-5a482fb51d44-var-run\") pod \"ovn-controller-ovs-dmbpm\" (UID: \"7f0d9af4-d6e3-4168-b83d-5a482fb51d44\") " pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.667238 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a-var-run\") pod \"ovn-controller-4vd4t\" (UID: \"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a\") " pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.667264 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfdl7\" (UniqueName: \"kubernetes.io/projected/79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a-kube-api-access-cfdl7\") pod \"ovn-controller-4vd4t\" (UID: \"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a\") " pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.667281 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdk86\" (UniqueName: \"kubernetes.io/projected/7f0d9af4-d6e3-4168-b83d-5a482fb51d44-kube-api-access-fdk86\") pod \"ovn-controller-ovs-dmbpm\" (UID: \"7f0d9af4-d6e3-4168-b83d-5a482fb51d44\") " pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.667327 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7f0d9af4-d6e3-4168-b83d-5a482fb51d44-var-log\") pod \"ovn-controller-ovs-dmbpm\" (UID: \"7f0d9af4-d6e3-4168-b83d-5a482fb51d44\") " pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.667496 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a-var-log-ovn\") pod \"ovn-controller-4vd4t\" (UID: \"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a\") " pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.667681 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7f0d9af4-d6e3-4168-b83d-5a482fb51d44-etc-ovs\") pod \"ovn-controller-ovs-dmbpm\" (UID: \"7f0d9af4-d6e3-4168-b83d-5a482fb51d44\") " pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.769381 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7f0d9af4-d6e3-4168-b83d-5a482fb51d44-etc-ovs\") pod \"ovn-controller-ovs-dmbpm\" (UID: \"7f0d9af4-d6e3-4168-b83d-5a482fb51d44\") " pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.769489 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a-combined-ca-bundle\") pod \"ovn-controller-4vd4t\" (UID: \"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a\") " pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.769519 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a-scripts\") pod \"ovn-controller-4vd4t\" (UID: \"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a\") " pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.769541 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7f0d9af4-d6e3-4168-b83d-5a482fb51d44-var-lib\") pod \"ovn-controller-ovs-dmbpm\" (UID: \"7f0d9af4-d6e3-4168-b83d-5a482fb51d44\") " pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.769578 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a-ovn-controller-tls-certs\") pod \"ovn-controller-4vd4t\" (UID: \"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a\") " pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.769604 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f0d9af4-d6e3-4168-b83d-5a482fb51d44-scripts\") pod \"ovn-controller-ovs-dmbpm\" (UID: \"7f0d9af4-d6e3-4168-b83d-5a482fb51d44\") " pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.769626 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a-var-run-ovn\") pod \"ovn-controller-4vd4t\" (UID: \"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a\") " pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.769648 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7f0d9af4-d6e3-4168-b83d-5a482fb51d44-var-run\") pod \"ovn-controller-ovs-dmbpm\" (UID: \"7f0d9af4-d6e3-4168-b83d-5a482fb51d44\") " pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.769675 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a-var-run\") pod \"ovn-controller-4vd4t\" (UID: \"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a\") " pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.769707 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfdl7\" (UniqueName: \"kubernetes.io/projected/79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a-kube-api-access-cfdl7\") pod \"ovn-controller-4vd4t\" (UID: \"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a\") " pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.769733 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdk86\" (UniqueName: \"kubernetes.io/projected/7f0d9af4-d6e3-4168-b83d-5a482fb51d44-kube-api-access-fdk86\") pod \"ovn-controller-ovs-dmbpm\" (UID: \"7f0d9af4-d6e3-4168-b83d-5a482fb51d44\") " pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.769796 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7f0d9af4-d6e3-4168-b83d-5a482fb51d44-var-log\") pod \"ovn-controller-ovs-dmbpm\" (UID: \"7f0d9af4-d6e3-4168-b83d-5a482fb51d44\") " pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.769849 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a-var-log-ovn\") pod \"ovn-controller-4vd4t\" (UID: \"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a\") " pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.771048 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7f0d9af4-d6e3-4168-b83d-5a482fb51d44-var-run\") pod \"ovn-controller-ovs-dmbpm\" (UID: \"7f0d9af4-d6e3-4168-b83d-5a482fb51d44\") " pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.771074 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a-var-run\") pod \"ovn-controller-4vd4t\" (UID: \"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a\") " pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.771102 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7f0d9af4-d6e3-4168-b83d-5a482fb51d44-var-log\") pod \"ovn-controller-ovs-dmbpm\" (UID: \"7f0d9af4-d6e3-4168-b83d-5a482fb51d44\") " pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.771129 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a-var-run-ovn\") pod \"ovn-controller-4vd4t\" (UID: \"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a\") " pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.771195 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7f0d9af4-d6e3-4168-b83d-5a482fb51d44-etc-ovs\") pod \"ovn-controller-ovs-dmbpm\" (UID: \"7f0d9af4-d6e3-4168-b83d-5a482fb51d44\") " pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.771233 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7f0d9af4-d6e3-4168-b83d-5a482fb51d44-var-lib\") pod \"ovn-controller-ovs-dmbpm\" (UID: \"7f0d9af4-d6e3-4168-b83d-5a482fb51d44\") " pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.771347 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a-var-log-ovn\") pod \"ovn-controller-4vd4t\" (UID: \"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a\") " pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.772421 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a-scripts\") pod \"ovn-controller-4vd4t\" (UID: \"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a\") " pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.774671 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f0d9af4-d6e3-4168-b83d-5a482fb51d44-scripts\") pod \"ovn-controller-ovs-dmbpm\" (UID: \"7f0d9af4-d6e3-4168-b83d-5a482fb51d44\") " pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.784182 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a-ovn-controller-tls-certs\") pod \"ovn-controller-4vd4t\" (UID: \"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a\") " pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.784336 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a-combined-ca-bundle\") pod \"ovn-controller-4vd4t\" (UID: \"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a\") " pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.788091 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfdl7\" (UniqueName: \"kubernetes.io/projected/79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a-kube-api-access-cfdl7\") pod \"ovn-controller-4vd4t\" (UID: \"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a\") " pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.790954 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdk86\" (UniqueName: \"kubernetes.io/projected/7f0d9af4-d6e3-4168-b83d-5a482fb51d44-kube-api-access-fdk86\") pod \"ovn-controller-ovs-dmbpm\" (UID: \"7f0d9af4-d6e3-4168-b83d-5a482fb51d44\") " pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.868907 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4vd4t" Mar 21 08:45:25 crc kubenswrapper[4696]: I0321 08:45:25.873433 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:27 crc kubenswrapper[4696]: W0321 08:45:27.295577 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ed4fc4e_1ec8_41a2_a8fb_d8fbc37071bc.slice/crio-7ef3c3ff0e08ca5b34710de32edf13933a61a288ca5029bbd1744dba9cdd986b WatchSource:0}: Error finding container 7ef3c3ff0e08ca5b34710de32edf13933a61a288ca5029bbd1744dba9cdd986b: Status 404 returned error can't find the container with id 7ef3c3ff0e08ca5b34710de32edf13933a61a288ca5029bbd1744dba9cdd986b Mar 21 08:45:27 crc kubenswrapper[4696]: I0321 08:45:27.819032 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 21 08:45:27 crc kubenswrapper[4696]: I0321 08:45:27.820691 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:27 crc kubenswrapper[4696]: I0321 08:45:27.822717 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Mar 21 08:45:27 crc kubenswrapper[4696]: I0321 08:45:27.823481 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Mar 21 08:45:27 crc kubenswrapper[4696]: I0321 08:45:27.823548 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Mar 21 08:45:27 crc kubenswrapper[4696]: I0321 08:45:27.823732 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Mar 21 08:45:27 crc kubenswrapper[4696]: I0321 08:45:27.823756 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-xvxmz" Mar 21 08:45:27 crc kubenswrapper[4696]: I0321 08:45:27.837064 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 21 08:45:27 crc kubenswrapper[4696]: I0321 08:45:27.872473 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" event={"ID":"4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc","Type":"ContainerStarted","Data":"7ef3c3ff0e08ca5b34710de32edf13933a61a288ca5029bbd1744dba9cdd986b"} Mar 21 08:45:27 crc kubenswrapper[4696]: I0321 08:45:27.908663 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9782617-ee6a-4ab5-aa9c-22dc382f654e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:27 crc kubenswrapper[4696]: I0321 08:45:27.908727 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ac321c5a-9c34-4307-b3ca-c213224c4459\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac321c5a-9c34-4307-b3ca-c213224c4459\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:27 crc kubenswrapper[4696]: I0321 08:45:27.908755 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9782617-ee6a-4ab5-aa9c-22dc382f654e-config\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:27 crc kubenswrapper[4696]: I0321 08:45:27.908795 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9782617-ee6a-4ab5-aa9c-22dc382f654e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:27 crc kubenswrapper[4696]: I0321 08:45:27.908838 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rs59\" (UniqueName: \"kubernetes.io/projected/b9782617-ee6a-4ab5-aa9c-22dc382f654e-kube-api-access-9rs59\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:27 crc kubenswrapper[4696]: I0321 08:45:27.908952 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9782617-ee6a-4ab5-aa9c-22dc382f654e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:27 crc kubenswrapper[4696]: I0321 08:45:27.909021 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b9782617-ee6a-4ab5-aa9c-22dc382f654e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:27 crc kubenswrapper[4696]: I0321 08:45:27.909049 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9782617-ee6a-4ab5-aa9c-22dc382f654e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.012413 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ac321c5a-9c34-4307-b3ca-c213224c4459\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac321c5a-9c34-4307-b3ca-c213224c4459\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.012466 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9782617-ee6a-4ab5-aa9c-22dc382f654e-config\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.012545 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9782617-ee6a-4ab5-aa9c-22dc382f654e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.012600 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rs59\" (UniqueName: \"kubernetes.io/projected/b9782617-ee6a-4ab5-aa9c-22dc382f654e-kube-api-access-9rs59\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.012643 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9782617-ee6a-4ab5-aa9c-22dc382f654e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.012691 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b9782617-ee6a-4ab5-aa9c-22dc382f654e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.012713 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9782617-ee6a-4ab5-aa9c-22dc382f654e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.012811 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9782617-ee6a-4ab5-aa9c-22dc382f654e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.013800 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9782617-ee6a-4ab5-aa9c-22dc382f654e-config\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.016902 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.016943 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ac321c5a-9c34-4307-b3ca-c213224c4459\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac321c5a-9c34-4307-b3ca-c213224c4459\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4c2671acde91ceefff4add7122203f4d9e4fa1a9877f8c9238d5f393a2679d43/globalmount\"" pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.021881 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9782617-ee6a-4ab5-aa9c-22dc382f654e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.023676 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b9782617-ee6a-4ab5-aa9c-22dc382f654e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.024147 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9782617-ee6a-4ab5-aa9c-22dc382f654e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.027119 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9782617-ee6a-4ab5-aa9c-22dc382f654e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.027130 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9782617-ee6a-4ab5-aa9c-22dc382f654e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.030854 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rs59\" (UniqueName: \"kubernetes.io/projected/b9782617-ee6a-4ab5-aa9c-22dc382f654e-kube-api-access-9rs59\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.057351 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ac321c5a-9c34-4307-b3ca-c213224c4459\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac321c5a-9c34-4307-b3ca-c213224c4459\") pod \"ovsdbserver-sb-0\" (UID: \"b9782617-ee6a-4ab5-aa9c-22dc382f654e\") " pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.138398 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.711100 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.714648 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.717769 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.717845 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-7fqqx" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.719095 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.719567 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.732396 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.826568 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21-config\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.826629 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-894pm\" (UniqueName: \"kubernetes.io/projected/2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21-kube-api-access-894pm\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.826684 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.826725 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.826781 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.826857 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.826882 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.826910 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-04756b01-0422-4538-8c5d-07718e628446\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-04756b01-0422-4538-8c5d-07718e628446\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.928750 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.929127 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.929292 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.929428 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.929528 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.929641 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-04756b01-0422-4538-8c5d-07718e628446\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-04756b01-0422-4538-8c5d-07718e628446\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.929783 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21-config\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.929931 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-894pm\" (UniqueName: \"kubernetes.io/projected/2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21-kube-api-access-894pm\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.930788 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.931278 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.931600 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21-config\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.934430 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.934467 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-04756b01-0422-4538-8c5d-07718e628446\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-04756b01-0422-4538-8c5d-07718e628446\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b68bed7f4b6bec78ebdd60dbc5855aee09bb29b8c09edb2a4b6a141eba87b9e5/globalmount\"" pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.943438 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.945766 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.946490 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.950116 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-894pm\" (UniqueName: \"kubernetes.io/projected/2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21-kube-api-access-894pm\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:28 crc kubenswrapper[4696]: I0321 08:45:28.966025 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-04756b01-0422-4538-8c5d-07718e628446\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-04756b01-0422-4538-8c5d-07718e628446\") pod \"ovsdbserver-nb-0\" (UID: \"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21\") " pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.034430 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.173145 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7"] Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.174120 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.176961 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-ca-bundle" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.177142 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-distributor-http" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.179803 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-config" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.180023 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-dockercfg-q29vw" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.186800 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-distributor-grpc" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.193462 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7"] Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.234612 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7\" (UID: \"a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7\") " pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.234869 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs2dk\" (UniqueName: \"kubernetes.io/projected/a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7-kube-api-access-rs2dk\") pod \"cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7\" (UID: \"a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7\") " pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.235049 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-distributor-http\" (UniqueName: \"kubernetes.io/secret/a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7-cloudkitty-lokistack-distributor-http\") pod \"cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7\" (UID: \"a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7\") " pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.235124 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7-config\") pod \"cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7\" (UID: \"a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7\") " pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.235151 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7-cloudkitty-lokistack-distributor-grpc\") pod \"cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7\" (UID: \"a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7\") " pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.336557 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs2dk\" (UniqueName: \"kubernetes.io/projected/a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7-kube-api-access-rs2dk\") pod \"cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7\" (UID: \"a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7\") " pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.336641 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-distributor-http\" (UniqueName: \"kubernetes.io/secret/a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7-cloudkitty-lokistack-distributor-http\") pod \"cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7\" (UID: \"a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7\") " pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.336678 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7-config\") pod \"cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7\" (UID: \"a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7\") " pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.336695 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7-cloudkitty-lokistack-distributor-grpc\") pod \"cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7\" (UID: \"a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7\") " pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.336728 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7\" (UID: \"a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7\") " pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.337645 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7\" (UID: \"a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7\") " pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.337973 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7-config\") pod \"cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7\" (UID: \"a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7\") " pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.339903 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-distributor-http\" (UniqueName: \"kubernetes.io/secret/a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7-cloudkitty-lokistack-distributor-http\") pod \"cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7\" (UID: \"a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7\") " pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.340309 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77"] Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.340763 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7-cloudkitty-lokistack-distributor-grpc\") pod \"cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7\" (UID: \"a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7\") " pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.341202 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.356786 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-querier-http" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.357040 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-querier-grpc" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.357085 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-loki-s3" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.366195 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs2dk\" (UniqueName: \"kubernetes.io/projected/a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7-kube-api-access-rs2dk\") pod \"cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7\" (UID: \"a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7\") " pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.378295 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77"] Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.438475 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/edf07829-775b-4672-89fe-81c3342300ae-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-querier-668f98fdd7-tnh77\" (UID: \"edf07829-775b-4672-89fe-81c3342300ae\") " pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.438547 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edf07829-775b-4672-89fe-81c3342300ae-config\") pod \"cloudkitty-lokistack-querier-668f98fdd7-tnh77\" (UID: \"edf07829-775b-4672-89fe-81c3342300ae\") " pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.438672 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-querier-grpc\" (UniqueName: \"kubernetes.io/secret/edf07829-775b-4672-89fe-81c3342300ae-cloudkitty-lokistack-querier-grpc\") pod \"cloudkitty-lokistack-querier-668f98fdd7-tnh77\" (UID: \"edf07829-775b-4672-89fe-81c3342300ae\") " pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.438730 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b87nc\" (UniqueName: \"kubernetes.io/projected/edf07829-775b-4672-89fe-81c3342300ae-kube-api-access-b87nc\") pod \"cloudkitty-lokistack-querier-668f98fdd7-tnh77\" (UID: \"edf07829-775b-4672-89fe-81c3342300ae\") " pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.438875 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/edf07829-775b-4672-89fe-81c3342300ae-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-querier-668f98fdd7-tnh77\" (UID: \"edf07829-775b-4672-89fe-81c3342300ae\") " pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.438986 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-querier-http\" (UniqueName: \"kubernetes.io/secret/edf07829-775b-4672-89fe-81c3342300ae-cloudkitty-lokistack-querier-http\") pod \"cloudkitty-lokistack-querier-668f98fdd7-tnh77\" (UID: \"edf07829-775b-4672-89fe-81c3342300ae\") " pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.493786 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.506396 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k"] Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.513436 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.524364 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-query-frontend-http" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.524558 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-query-frontend-grpc" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.529998 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k"] Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.540537 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/edf07829-775b-4672-89fe-81c3342300ae-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-querier-668f98fdd7-tnh77\" (UID: \"edf07829-775b-4672-89fe-81c3342300ae\") " pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.540837 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/bbf61ae0-6918-44af-be69-ac8a220fcd6b-cloudkitty-lokistack-query-frontend-http\") pod \"cloudkitty-lokistack-query-frontend-6f54889599-6mn7k\" (UID: \"bbf61ae0-6918-44af-be69-ac8a220fcd6b\") " pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.542671 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/edf07829-775b-4672-89fe-81c3342300ae-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-querier-668f98fdd7-tnh77\" (UID: \"edf07829-775b-4672-89fe-81c3342300ae\") " pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.549750 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtg8g\" (UniqueName: \"kubernetes.io/projected/bbf61ae0-6918-44af-be69-ac8a220fcd6b-kube-api-access-dtg8g\") pod \"cloudkitty-lokistack-query-frontend-6f54889599-6mn7k\" (UID: \"bbf61ae0-6918-44af-be69-ac8a220fcd6b\") " pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.549794 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-querier-http\" (UniqueName: \"kubernetes.io/secret/edf07829-775b-4672-89fe-81c3342300ae-cloudkitty-lokistack-querier-http\") pod \"cloudkitty-lokistack-querier-668f98fdd7-tnh77\" (UID: \"edf07829-775b-4672-89fe-81c3342300ae\") " pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.549837 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbf61ae0-6918-44af-be69-ac8a220fcd6b-config\") pod \"cloudkitty-lokistack-query-frontend-6f54889599-6mn7k\" (UID: \"bbf61ae0-6918-44af-be69-ac8a220fcd6b\") " pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.549861 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/bbf61ae0-6918-44af-be69-ac8a220fcd6b-cloudkitty-lokistack-query-frontend-grpc\") pod \"cloudkitty-lokistack-query-frontend-6f54889599-6mn7k\" (UID: \"bbf61ae0-6918-44af-be69-ac8a220fcd6b\") " pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.549886 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/edf07829-775b-4672-89fe-81c3342300ae-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-querier-668f98fdd7-tnh77\" (UID: \"edf07829-775b-4672-89fe-81c3342300ae\") " pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.549939 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edf07829-775b-4672-89fe-81c3342300ae-config\") pod \"cloudkitty-lokistack-querier-668f98fdd7-tnh77\" (UID: \"edf07829-775b-4672-89fe-81c3342300ae\") " pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.549970 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-querier-grpc\" (UniqueName: \"kubernetes.io/secret/edf07829-775b-4672-89fe-81c3342300ae-cloudkitty-lokistack-querier-grpc\") pod \"cloudkitty-lokistack-querier-668f98fdd7-tnh77\" (UID: \"edf07829-775b-4672-89fe-81c3342300ae\") " pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.549990 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bbf61ae0-6918-44af-be69-ac8a220fcd6b-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-query-frontend-6f54889599-6mn7k\" (UID: \"bbf61ae0-6918-44af-be69-ac8a220fcd6b\") " pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.550032 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b87nc\" (UniqueName: \"kubernetes.io/projected/edf07829-775b-4672-89fe-81c3342300ae-kube-api-access-b87nc\") pod \"cloudkitty-lokistack-querier-668f98fdd7-tnh77\" (UID: \"edf07829-775b-4672-89fe-81c3342300ae\") " pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.552149 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edf07829-775b-4672-89fe-81c3342300ae-config\") pod \"cloudkitty-lokistack-querier-668f98fdd7-tnh77\" (UID: \"edf07829-775b-4672-89fe-81c3342300ae\") " pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.553250 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/edf07829-775b-4672-89fe-81c3342300ae-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-querier-668f98fdd7-tnh77\" (UID: \"edf07829-775b-4672-89fe-81c3342300ae\") " pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.553351 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-querier-http\" (UniqueName: \"kubernetes.io/secret/edf07829-775b-4672-89fe-81c3342300ae-cloudkitty-lokistack-querier-http\") pod \"cloudkitty-lokistack-querier-668f98fdd7-tnh77\" (UID: \"edf07829-775b-4672-89fe-81c3342300ae\") " pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.557575 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-querier-grpc\" (UniqueName: \"kubernetes.io/secret/edf07829-775b-4672-89fe-81c3342300ae-cloudkitty-lokistack-querier-grpc\") pod \"cloudkitty-lokistack-querier-668f98fdd7-tnh77\" (UID: \"edf07829-775b-4672-89fe-81c3342300ae\") " pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.571202 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b87nc\" (UniqueName: \"kubernetes.io/projected/edf07829-775b-4672-89fe-81c3342300ae-kube-api-access-b87nc\") pod \"cloudkitty-lokistack-querier-668f98fdd7-tnh77\" (UID: \"edf07829-775b-4672-89fe-81c3342300ae\") " pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.608173 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w"] Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.609419 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.616809 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w"] Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.619845 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-gateway-ca-bundle" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.620051 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway-client-http" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.620153 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-gateway" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.620556 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway-http" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.620644 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-ca" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.620722 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.630227 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8"] Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.631685 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.633799 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway-dockercfg-dnjkl" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.643766 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8"] Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.651534 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/70a97941-5dea-48a7-8dbf-b475b4af0b6e-tls-secret\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.651571 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70a97941-5dea-48a7-8dbf-b475b4af0b6e-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.651718 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/70a97941-5dea-48a7-8dbf-b475b4af0b6e-rbac\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.651840 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/bbf61ae0-6918-44af-be69-ac8a220fcd6b-cloudkitty-lokistack-query-frontend-http\") pod \"cloudkitty-lokistack-query-frontend-6f54889599-6mn7k\" (UID: \"bbf61ae0-6918-44af-be69-ac8a220fcd6b\") " pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.651925 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/70a97941-5dea-48a7-8dbf-b475b4af0b6e-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.651949 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/70a97941-5dea-48a7-8dbf-b475b4af0b6e-tenants\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.651978 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/70a97941-5dea-48a7-8dbf-b475b4af0b6e-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.652076 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtg8g\" (UniqueName: \"kubernetes.io/projected/bbf61ae0-6918-44af-be69-ac8a220fcd6b-kube-api-access-dtg8g\") pod \"cloudkitty-lokistack-query-frontend-6f54889599-6mn7k\" (UID: \"bbf61ae0-6918-44af-be69-ac8a220fcd6b\") " pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.652114 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbf61ae0-6918-44af-be69-ac8a220fcd6b-config\") pod \"cloudkitty-lokistack-query-frontend-6f54889599-6mn7k\" (UID: \"bbf61ae0-6918-44af-be69-ac8a220fcd6b\") " pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.652151 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/bbf61ae0-6918-44af-be69-ac8a220fcd6b-cloudkitty-lokistack-query-frontend-grpc\") pod \"cloudkitty-lokistack-query-frontend-6f54889599-6mn7k\" (UID: \"bbf61ae0-6918-44af-be69-ac8a220fcd6b\") " pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.652170 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70a97941-5dea-48a7-8dbf-b475b4af0b6e-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.652277 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bbf61ae0-6918-44af-be69-ac8a220fcd6b-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-query-frontend-6f54889599-6mn7k\" (UID: \"bbf61ae0-6918-44af-be69-ac8a220fcd6b\") " pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.652336 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70a97941-5dea-48a7-8dbf-b475b4af0b6e-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.652403 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dzgs\" (UniqueName: \"kubernetes.io/projected/70a97941-5dea-48a7-8dbf-b475b4af0b6e-kube-api-access-9dzgs\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.654509 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbf61ae0-6918-44af-be69-ac8a220fcd6b-config\") pod \"cloudkitty-lokistack-query-frontend-6f54889599-6mn7k\" (UID: \"bbf61ae0-6918-44af-be69-ac8a220fcd6b\") " pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.654755 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bbf61ae0-6918-44af-be69-ac8a220fcd6b-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-query-frontend-6f54889599-6mn7k\" (UID: \"bbf61ae0-6918-44af-be69-ac8a220fcd6b\") " pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.657571 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/bbf61ae0-6918-44af-be69-ac8a220fcd6b-cloudkitty-lokistack-query-frontend-grpc\") pod \"cloudkitty-lokistack-query-frontend-6f54889599-6mn7k\" (UID: \"bbf61ae0-6918-44af-be69-ac8a220fcd6b\") " pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.657612 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/bbf61ae0-6918-44af-be69-ac8a220fcd6b-cloudkitty-lokistack-query-frontend-http\") pod \"cloudkitty-lokistack-query-frontend-6f54889599-6mn7k\" (UID: \"bbf61ae0-6918-44af-be69-ac8a220fcd6b\") " pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.672356 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtg8g\" (UniqueName: \"kubernetes.io/projected/bbf61ae0-6918-44af-be69-ac8a220fcd6b-kube-api-access-dtg8g\") pod \"cloudkitty-lokistack-query-frontend-6f54889599-6mn7k\" (UID: \"bbf61ae0-6918-44af-be69-ac8a220fcd6b\") " pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.725327 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.755924 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/70a97941-5dea-48a7-8dbf-b475b4af0b6e-tls-secret\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.755963 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70a97941-5dea-48a7-8dbf-b475b4af0b6e-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.756006 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/70a97941-5dea-48a7-8dbf-b475b4af0b6e-rbac\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.756049 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/70a97941-5dea-48a7-8dbf-b475b4af0b6e-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.756084 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/70a97941-5dea-48a7-8dbf-b475b4af0b6e-tenants\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.756105 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/70a97941-5dea-48a7-8dbf-b475b4af0b6e-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.756162 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-tenants\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.756177 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-tls-secret\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.756192 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.756227 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70a97941-5dea-48a7-8dbf-b475b4af0b6e-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.756246 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.756275 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgmrw\" (UniqueName: \"kubernetes.io/projected/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-kube-api-access-kgmrw\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.756331 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.756354 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.756391 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.756420 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70a97941-5dea-48a7-8dbf-b475b4af0b6e-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.756435 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-rbac\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.756481 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dzgs\" (UniqueName: \"kubernetes.io/projected/70a97941-5dea-48a7-8dbf-b475b4af0b6e-kube-api-access-9dzgs\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.757115 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/70a97941-5dea-48a7-8dbf-b475b4af0b6e-rbac\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.757164 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/70a97941-5dea-48a7-8dbf-b475b4af0b6e-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: E0321 08:45:29.757243 4696 secret.go:188] Couldn't get secret openstack/cloudkitty-lokistack-gateway-http: secret "cloudkitty-lokistack-gateway-http" not found Mar 21 08:45:29 crc kubenswrapper[4696]: E0321 08:45:29.757286 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/70a97941-5dea-48a7-8dbf-b475b4af0b6e-tls-secret podName:70a97941-5dea-48a7-8dbf-b475b4af0b6e nodeName:}" failed. No retries permitted until 2026-03-21 08:45:30.257271569 +0000 UTC m=+1064.378152282 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/70a97941-5dea-48a7-8dbf-b475b4af0b6e-tls-secret") pod "cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" (UID: "70a97941-5dea-48a7-8dbf-b475b4af0b6e") : secret "cloudkitty-lokistack-gateway-http" not found Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.758511 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70a97941-5dea-48a7-8dbf-b475b4af0b6e-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.758636 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70a97941-5dea-48a7-8dbf-b475b4af0b6e-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.758525 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70a97941-5dea-48a7-8dbf-b475b4af0b6e-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.759158 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/70a97941-5dea-48a7-8dbf-b475b4af0b6e-tenants\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.760616 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/70a97941-5dea-48a7-8dbf-b475b4af0b6e-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.770499 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dzgs\" (UniqueName: \"kubernetes.io/projected/70a97941-5dea-48a7-8dbf-b475b4af0b6e-kube-api-access-9dzgs\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.849672 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.857948 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-tls-secret\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.857988 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.858007 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-tenants\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.858029 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.858048 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgmrw\" (UniqueName: \"kubernetes.io/projected/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-kube-api-access-kgmrw\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.858077 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.858101 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.858121 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.858148 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-rbac\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.859114 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-rbac\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.859709 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.859894 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.859961 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.860120 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.861540 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-tls-secret\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.862024 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.863351 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-tenants\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.881549 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgmrw\" (UniqueName: \"kubernetes.io/projected/d2b0b7e1-1fa0-44ce-bab0-0964460e645b-kube-api-access-kgmrw\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-lczl8\" (UID: \"d2b0b7e1-1fa0-44ce-bab0-0964460e645b\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:29 crc kubenswrapper[4696]: I0321 08:45:29.951015 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.264785 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/70a97941-5dea-48a7-8dbf-b475b4af0b6e-tls-secret\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.282436 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/70a97941-5dea-48a7-8dbf-b475b4af0b6e-tls-secret\") pod \"cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w\" (UID: \"70a97941-5dea-48a7-8dbf-b475b4af0b6e\") " pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.322859 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-ingester-0"] Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.324012 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.327712 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-ingester-grpc" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.327939 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-ingester-http" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.334256 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-ingester-0"] Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.366148 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d6518eba-34ec-4a38-b279-cd8bb11e7e24-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.366188 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/d6518eba-34ec-4a38-b279-cd8bb11e7e24-cloudkitty-lokistack-ingester-grpc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.366248 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/d6518eba-34ec-4a38-b279-cd8bb11e7e24-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.366285 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.366308 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6518eba-34ec-4a38-b279-cd8bb11e7e24-config\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.366333 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ingester-http\" (UniqueName: \"kubernetes.io/secret/d6518eba-34ec-4a38-b279-cd8bb11e7e24-cloudkitty-lokistack-ingester-http\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.366357 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pv4vs\" (UniqueName: \"kubernetes.io/projected/d6518eba-34ec-4a38-b279-cd8bb11e7e24-kube-api-access-pv4vs\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.366380 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.430431 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-compactor-0"] Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.431624 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.433608 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-compactor-http" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.433758 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-compactor-grpc" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.446003 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-compactor-0"] Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.475478 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6518eba-34ec-4a38-b279-cd8bb11e7e24-config\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.475537 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ingester-http\" (UniqueName: \"kubernetes.io/secret/d6518eba-34ec-4a38-b279-cd8bb11e7e24-cloudkitty-lokistack-ingester-http\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.475607 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pv4vs\" (UniqueName: \"kubernetes.io/projected/d6518eba-34ec-4a38-b279-cd8bb11e7e24-kube-api-access-pv4vs\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.475639 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.475703 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d6518eba-34ec-4a38-b279-cd8bb11e7e24-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.475744 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/d6518eba-34ec-4a38-b279-cd8bb11e7e24-cloudkitty-lokistack-ingester-grpc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.475839 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/d6518eba-34ec-4a38-b279-cd8bb11e7e24-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.475897 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.476251 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.476256 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.478787 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6518eba-34ec-4a38-b279-cd8bb11e7e24-config\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.478949 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d6518eba-34ec-4a38-b279-cd8bb11e7e24-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.487385 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ingester-http\" (UniqueName: \"kubernetes.io/secret/d6518eba-34ec-4a38-b279-cd8bb11e7e24-cloudkitty-lokistack-ingester-http\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.488066 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/d6518eba-34ec-4a38-b279-cd8bb11e7e24-cloudkitty-lokistack-ingester-grpc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.488066 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/d6518eba-34ec-4a38-b279-cd8bb11e7e24-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.505549 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pv4vs\" (UniqueName: \"kubernetes.io/projected/d6518eba-34ec-4a38-b279-cd8bb11e7e24-kube-api-access-pv4vs\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.506964 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.510944 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"d6518eba-34ec-4a38-b279-cd8bb11e7e24\") " pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.535732 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.546785 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-index-gateway-0"] Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.548633 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.561475 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-index-gateway-http" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.561751 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-index-gateway-grpc" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.567028 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-index-gateway-0"] Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.582071 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") " pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.582392 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8-config\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") " pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.582946 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") " pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.583007 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8-cloudkitty-lokistack-compactor-grpc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") " pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.583039 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lktqs\" (UniqueName: \"kubernetes.io/projected/14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8-kube-api-access-lktqs\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") " pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.583099 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-compactor-http\" (UniqueName: \"kubernetes.io/secret/14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8-cloudkitty-lokistack-compactor-http\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") " pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.583128 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") " pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.652578 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.689677 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88865f3d-9bb9-4844-bb96-dc25c11615ec-config\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.689825 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8-cloudkitty-lokistack-compactor-grpc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") " pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.689860 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lktqs\" (UniqueName: \"kubernetes.io/projected/14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8-kube-api-access-lktqs\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") " pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.689888 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs4kd\" (UniqueName: \"kubernetes.io/projected/88865f3d-9bb9-4844-bb96-dc25c11615ec-kube-api-access-xs4kd\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.689922 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/88865f3d-9bb9-4844-bb96-dc25c11615ec-cloudkitty-lokistack-index-gateway-grpc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.689941 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88865f3d-9bb9-4844-bb96-dc25c11615ec-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.690200 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-compactor-http\" (UniqueName: \"kubernetes.io/secret/14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8-cloudkitty-lokistack-compactor-http\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") " pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.690294 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") " pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.690386 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.690525 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") " pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.690584 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/88865f3d-9bb9-4844-bb96-dc25c11615ec-cloudkitty-lokistack-index-gateway-http\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.690710 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/88865f3d-9bb9-4844-bb96-dc25c11615ec-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.690747 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8-config\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") " pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.690843 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") " pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.691015 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.692258 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") " pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.692799 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8-config\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") " pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.698387 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-compactor-http\" (UniqueName: \"kubernetes.io/secret/14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8-cloudkitty-lokistack-compactor-http\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") " pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.702842 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") " pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.703662 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8-cloudkitty-lokistack-compactor-grpc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") " pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.711250 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lktqs\" (UniqueName: \"kubernetes.io/projected/14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8-kube-api-access-lktqs\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") " pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.720095 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8\") " pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.792461 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88865f3d-9bb9-4844-bb96-dc25c11615ec-config\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.792542 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs4kd\" (UniqueName: \"kubernetes.io/projected/88865f3d-9bb9-4844-bb96-dc25c11615ec-kube-api-access-xs4kd\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.792591 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/88865f3d-9bb9-4844-bb96-dc25c11615ec-cloudkitty-lokistack-index-gateway-grpc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.792622 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88865f3d-9bb9-4844-bb96-dc25c11615ec-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.792675 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.792744 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/88865f3d-9bb9-4844-bb96-dc25c11615ec-cloudkitty-lokistack-index-gateway-http\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.792788 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/88865f3d-9bb9-4844-bb96-dc25c11615ec-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.793196 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.794140 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88865f3d-9bb9-4844-bb96-dc25c11615ec-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.794449 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88865f3d-9bb9-4844-bb96-dc25c11615ec-config\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.796445 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/88865f3d-9bb9-4844-bb96-dc25c11615ec-cloudkitty-lokistack-index-gateway-grpc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.796959 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/88865f3d-9bb9-4844-bb96-dc25c11615ec-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.797356 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/88865f3d-9bb9-4844-bb96-dc25c11615ec-cloudkitty-lokistack-index-gateway-http\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.815895 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.820746 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs4kd\" (UniqueName: \"kubernetes.io/projected/88865f3d-9bb9-4844-bb96-dc25c11615ec-kube-api-access-xs4kd\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"88865f3d-9bb9-4844-bb96-dc25c11615ec\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.862593 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:45:30 crc kubenswrapper[4696]: I0321 08:45:30.880601 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:33 crc kubenswrapper[4696]: E0321 08:45:33.413937 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 21 08:45:33 crc kubenswrapper[4696]: E0321 08:45:33.414133 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ccz2l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-4xhpm_openstack(c31577bd-23ed-4ab0-8cf1-42a40b0cc941): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 08:45:33 crc kubenswrapper[4696]: E0321 08:45:33.416986 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-4xhpm" podUID="c31577bd-23ed-4ab0-8cf1-42a40b0cc941" Mar 21 08:45:33 crc kubenswrapper[4696]: E0321 08:45:33.438283 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 21 08:45:33 crc kubenswrapper[4696]: E0321 08:45:33.438697 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nrk4w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-hxprq_openstack(26009008-54e8-48fc-b1b5-99c0e0c1e544): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 08:45:33 crc kubenswrapper[4696]: E0321 08:45:33.439895 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-hxprq" podUID="26009008-54e8-48fc-b1b5-99c0e0c1e544" Mar 21 08:45:33 crc kubenswrapper[4696]: I0321 08:45:33.897933 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 21 08:45:34 crc kubenswrapper[4696]: I0321 08:45:34.089055 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 21 08:45:34 crc kubenswrapper[4696]: I0321 08:45:34.888282 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-hxprq" Mar 21 08:45:34 crc kubenswrapper[4696]: I0321 08:45:34.910480 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-4xhpm" Mar 21 08:45:34 crc kubenswrapper[4696]: I0321 08:45:34.935900 4696 generic.go:334] "Generic (PLEG): container finished" podID="4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc" containerID="de22920663ad6577ced117df929d32519b87df6661cd629b2f26fe803f1f0cc8" exitCode=0 Mar 21 08:45:34 crc kubenswrapper[4696]: I0321 08:45:34.935964 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" event={"ID":"4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc","Type":"ContainerDied","Data":"de22920663ad6577ced117df929d32519b87df6661cd629b2f26fe803f1f0cc8"} Mar 21 08:45:34 crc kubenswrapper[4696]: I0321 08:45:34.944134 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-4xhpm" event={"ID":"c31577bd-23ed-4ab0-8cf1-42a40b0cc941","Type":"ContainerDied","Data":"9084f0474616d92968667f58dd6404e15d4363745f600cb2f5e1805478d5a25e"} Mar 21 08:45:34 crc kubenswrapper[4696]: I0321 08:45:34.944120 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-4xhpm" Mar 21 08:45:34 crc kubenswrapper[4696]: I0321 08:45:34.953428 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b057ad21-8030-49e9-b400-c36d433b9f8c","Type":"ContainerStarted","Data":"8c465eaf26c5aa0886cd99d3f25096099b9e93a7f62b3cae2b59029502c9e221"} Mar 21 08:45:34 crc kubenswrapper[4696]: I0321 08:45:34.962150 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"919a11c7-cf69-4660-8de6-6b1a54b63f26","Type":"ContainerStarted","Data":"2286e94d43497d5d4c2be41ba5afa258a5f604d5a2b53d8e50432565f39fc5c8"} Mar 21 08:45:34 crc kubenswrapper[4696]: I0321 08:45:34.977889 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77"] Mar 21 08:45:34 crc kubenswrapper[4696]: I0321 08:45:34.983047 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26009008-54e8-48fc-b1b5-99c0e0c1e544-dns-svc\") pod \"26009008-54e8-48fc-b1b5-99c0e0c1e544\" (UID: \"26009008-54e8-48fc-b1b5-99c0e0c1e544\") " Mar 21 08:45:34 crc kubenswrapper[4696]: I0321 08:45:34.983092 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26009008-54e8-48fc-b1b5-99c0e0c1e544-config\") pod \"26009008-54e8-48fc-b1b5-99c0e0c1e544\" (UID: \"26009008-54e8-48fc-b1b5-99c0e0c1e544\") " Mar 21 08:45:34 crc kubenswrapper[4696]: I0321 08:45:34.983132 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrk4w\" (UniqueName: \"kubernetes.io/projected/26009008-54e8-48fc-b1b5-99c0e0c1e544-kube-api-access-nrk4w\") pod \"26009008-54e8-48fc-b1b5-99c0e0c1e544\" (UID: \"26009008-54e8-48fc-b1b5-99c0e0c1e544\") " Mar 21 08:45:34 crc kubenswrapper[4696]: I0321 08:45:34.985494 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26009008-54e8-48fc-b1b5-99c0e0c1e544-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "26009008-54e8-48fc-b1b5-99c0e0c1e544" (UID: "26009008-54e8-48fc-b1b5-99c0e0c1e544"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:45:34 crc kubenswrapper[4696]: I0321 08:45:34.985877 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26009008-54e8-48fc-b1b5-99c0e0c1e544-config" (OuterVolumeSpecName: "config") pod "26009008-54e8-48fc-b1b5-99c0e0c1e544" (UID: "26009008-54e8-48fc-b1b5-99c0e0c1e544"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:45:34 crc kubenswrapper[4696]: I0321 08:45:34.986396 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-hxprq" event={"ID":"26009008-54e8-48fc-b1b5-99c0e0c1e544","Type":"ContainerDied","Data":"d19214f1fe3e7145632c88ea346e1722182855d39c172586721d90f457105b56"} Mar 21 08:45:34 crc kubenswrapper[4696]: I0321 08:45:34.987033 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-hxprq" Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:34.997425 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26009008-54e8-48fc-b1b5-99c0e0c1e544-kube-api-access-nrk4w" (OuterVolumeSpecName: "kube-api-access-nrk4w") pod "26009008-54e8-48fc-b1b5-99c0e0c1e544" (UID: "26009008-54e8-48fc-b1b5-99c0e0c1e544"). InnerVolumeSpecName "kube-api-access-nrk4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:45:35 crc kubenswrapper[4696]: W0321 08:45:35.001247 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1b37b82_bd6a_4e10_98cd_89fbff6bb554.slice/crio-4fb701a4fed069904e046a153906aef0ff0e084a70050acd1db25ff37cc27178 WatchSource:0}: Error finding container 4fb701a4fed069904e046a153906aef0ff0e084a70050acd1db25ff37cc27178: Status 404 returned error can't find the container with id 4fb701a4fed069904e046a153906aef0ff0e084a70050acd1db25ff37cc27178 Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.001360 4696 generic.go:334] "Generic (PLEG): container finished" podID="065e4107-c490-498d-a0f0-fce4011c38f1" containerID="a85da9b90c784172d64500bb8c3d9906079ca9382be826d2a0cc79912eede670" exitCode=0 Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.001388 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" event={"ID":"065e4107-c490-498d-a0f0-fce4011c38f1","Type":"ContainerDied","Data":"a85da9b90c784172d64500bb8c3d9906079ca9382be826d2a0cc79912eede670"} Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.050548 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Mar 21 08:45:35 crc kubenswrapper[4696]: W0321 08:45:35.081334 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ea6ac8c_cb0e_46a5_b3a2_0701bd7dce0a.slice/crio-589d42286ee21cc59f6a9512729069f2eff8a5e35c1b64b7c8364d3eb6a79e09 WatchSource:0}: Error finding container 589d42286ee21cc59f6a9512729069f2eff8a5e35c1b64b7c8364d3eb6a79e09: Status 404 returned error can't find the container with id 589d42286ee21cc59f6a9512729069f2eff8a5e35c1b64b7c8364d3eb6a79e09 Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.084322 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31577bd-23ed-4ab0-8cf1-42a40b0cc941-config\") pod \"c31577bd-23ed-4ab0-8cf1-42a40b0cc941\" (UID: \"c31577bd-23ed-4ab0-8cf1-42a40b0cc941\") " Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.084425 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccz2l\" (UniqueName: \"kubernetes.io/projected/c31577bd-23ed-4ab0-8cf1-42a40b0cc941-kube-api-access-ccz2l\") pod \"c31577bd-23ed-4ab0-8cf1-42a40b0cc941\" (UID: \"c31577bd-23ed-4ab0-8cf1-42a40b0cc941\") " Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.085407 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c31577bd-23ed-4ab0-8cf1-42a40b0cc941-config" (OuterVolumeSpecName: "config") pod "c31577bd-23ed-4ab0-8cf1-42a40b0cc941" (UID: "c31577bd-23ed-4ab0-8cf1-42a40b0cc941"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.085550 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26009008-54e8-48fc-b1b5-99c0e0c1e544-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.085939 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26009008-54e8-48fc-b1b5-99c0e0c1e544-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.086661 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrk4w\" (UniqueName: \"kubernetes.io/projected/26009008-54e8-48fc-b1b5-99c0e0c1e544-kube-api-access-nrk4w\") on node \"crc\" DevicePath \"\"" Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.090366 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c31577bd-23ed-4ab0-8cf1-42a40b0cc941-kube-api-access-ccz2l" (OuterVolumeSpecName: "kube-api-access-ccz2l") pod "c31577bd-23ed-4ab0-8cf1-42a40b0cc941" (UID: "c31577bd-23ed-4ab0-8cf1-42a40b0cc941"). InnerVolumeSpecName "kube-api-access-ccz2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.124565 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.134796 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.152749 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.188633 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31577bd-23ed-4ab0-8cf1-42a40b0cc941-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.188665 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccz2l\" (UniqueName: \"kubernetes.io/projected/c31577bd-23ed-4ab0-8cf1-42a40b0cc941-kube-api-access-ccz2l\") on node \"crc\" DevicePath \"\"" Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.202088 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4vd4t"] Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.208378 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 21 08:45:35 crc kubenswrapper[4696]: E0321 08:45:35.291932 4696 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Mar 21 08:45:35 crc kubenswrapper[4696]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/065e4107-c490-498d-a0f0-fce4011c38f1/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Mar 21 08:45:35 crc kubenswrapper[4696]: > podSandboxID="32e469f0a3f6c7870fde7ab3d4bd0355051fe89b9f5755da50661bd7b0af38fd" Mar 21 08:45:35 crc kubenswrapper[4696]: E0321 08:45:35.292114 4696 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 21 08:45:35 crc kubenswrapper[4696]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tnr7n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5ccc8479f9-27j2f_openstack(065e4107-c490-498d-a0f0-fce4011c38f1): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/065e4107-c490-498d-a0f0-fce4011c38f1/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Mar 21 08:45:35 crc kubenswrapper[4696]: > logger="UnhandledError" Mar 21 08:45:35 crc kubenswrapper[4696]: E0321 08:45:35.293651 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/065e4107-c490-498d-a0f0-fce4011c38f1/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" podUID="065e4107-c490-498d-a0f0-fce4011c38f1" Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.385223 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4xhpm"] Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.392062 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4xhpm"] Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.432958 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-hxprq"] Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.444224 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-hxprq"] Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.511997 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8"] Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.520869 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-index-gateway-0"] Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.542147 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-ingester-0"] Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.559889 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-compactor-0"] Mar 21 08:45:35 crc kubenswrapper[4696]: W0321 08:45:35.576177 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2b0b7e1_1fa0_44ce_bab0_0964460e645b.slice/crio-e85174d41740fe6580a0b0c138bf1d43cc6eb092ec564512e406313b7259c701 WatchSource:0}: Error finding container e85174d41740fe6580a0b0c138bf1d43cc6eb092ec564512e406313b7259c701: Status 404 returned error can't find the container with id e85174d41740fe6580a0b0c138bf1d43cc6eb092ec564512e406313b7259c701 Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.576282 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7"] Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.582897 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k"] Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.589558 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.604636 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w"] Mar 21 08:45:35 crc kubenswrapper[4696]: W0321 08:45:35.610278 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14c7fb5d_c3dd_4751_b2a2_b4cc618cd7a8.slice/crio-114545dc83ec20e038a21c521658598dd0ce4a9e9fcfc5b2b2fe1ea1ca836c47 WatchSource:0}: Error finding container 114545dc83ec20e038a21c521658598dd0ce4a9e9fcfc5b2b2fe1ea1ca836c47: Status 404 returned error can't find the container with id 114545dc83ec20e038a21c521658598dd0ce4a9e9fcfc5b2b2fe1ea1ca836c47 Mar 21 08:45:35 crc kubenswrapper[4696]: W0321 08:45:35.619780 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b9177a1_3d3a_4a87_a1fa_1dbe96439288.slice/crio-fa2c7c7d410cc660805060a6002d5068a2194c994a8895b165f06b0b23a313d7 WatchSource:0}: Error finding container fa2c7c7d410cc660805060a6002d5068a2194c994a8895b165f06b0b23a313d7: Status 404 returned error can't find the container with id fa2c7c7d410cc660805060a6002d5068a2194c994a8895b165f06b0b23a313d7 Mar 21 08:45:35 crc kubenswrapper[4696]: E0321 08:45:35.624784 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:gateway,Image:registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:86738dca2db56373b411adb51a4dfde274968f8c8fde42dbc2daf4bac3bb8daf,Command:[],Args:[--debug.name=lokistack-gateway --web.listen=0.0.0.0:8080 --web.internal.listen=0.0.0.0:8081 --web.healthchecks.url=https://localhost:8080 --log.level=warn --logs.read.endpoint=https://cloudkitty-lokistack-query-frontend-http.openstack.svc.cluster.local:3100 --logs.tail.endpoint=https://cloudkitty-lokistack-query-frontend-http.openstack.svc.cluster.local:3100 --logs.write.endpoint=https://cloudkitty-lokistack-distributor-http.openstack.svc.cluster.local:3100 --logs.write-timeout=4m0s --rbac.config=/etc/lokistack-gateway/rbac.yaml --tenants.config=/etc/lokistack-gateway/tenants.yaml --server.read-timeout=48s --server.write-timeout=6m0s --tls.min-version=VersionTLS12 --tls.server.cert-file=/var/run/tls/http/server/tls.crt --tls.server.key-file=/var/run/tls/http/server/tls.key --tls.healthchecks.server-ca-file=/var/run/ca/server/service-ca.crt --tls.healthchecks.server-name=cloudkitty-lokistack-gateway-http.openstack.svc.cluster.local --tls.internal.server.cert-file=/var/run/tls/http/server/tls.crt --tls.internal.server.key-file=/var/run/tls/http/server/tls.key --tls.min-version=VersionTLS12 --tls.cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --logs.tls.ca-file=/var/run/ca/upstream/service-ca.crt --logs.tls.cert-file=/var/run/tls/http/upstream/tls.crt --logs.tls.key-file=/var/run/tls/http/upstream/tls.key --tls.client-auth-type=RequestClientCert],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},ContainerPort{Name:public,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:rbac,ReadOnly:true,MountPath:/etc/lokistack-gateway/rbac.yaml,SubPath:rbac.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tenants,ReadOnly:true,MountPath:/etc/lokistack-gateway/tenants.yaml,SubPath:tenants.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:lokistack-gateway,ReadOnly:true,MountPath:/etc/lokistack-gateway/lokistack-gateway.rego,SubPath:lokistack-gateway.rego,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tls-secret,ReadOnly:true,MountPath:/var/run/tls/http/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-gateway-client-http,ReadOnly:true,MountPath:/var/run/tls/http/upstream,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-ca-bundle,ReadOnly:true,MountPath:/var/run/ca/upstream,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-gateway-ca-bundle,ReadOnly:true,MountPath:/var/run/ca/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-ca-bundle,ReadOnly:false,MountPath:/var/run/tenants-ca/cloudkitty,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9dzgs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/live,Port:{0 8081 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:2,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/ready,Port:{0 8081 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:12,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w_openstack(70a97941-5dea-48a7-8dbf-b475b4af0b6e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 21 08:45:35 crc kubenswrapper[4696]: E0321 08:45:35.625963 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" podUID="70a97941-5dea-48a7-8dbf-b475b4af0b6e" Mar 21 08:45:35 crc kubenswrapper[4696]: E0321 08:45:35.666326 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init-config-reloader,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:e4412f5688c9725f36d2f566f624d82a1a2a5b957686245fd2defcc39604bdc2,Command:[/bin/prometheus-config-reloader],Args:[--watch-interval=0 --listen-address=:8081 --config-file=/etc/prometheus/config/prometheus.yaml.gz --config-envsubst-file=/etc/prometheus/config_out/prometheus.env.yaml --watched-dir=/etc/prometheus/rules/prometheus-metric-storage-rulefiles-0 --watched-dir=/etc/prometheus/rules/prometheus-metric-storage-rulefiles-1 --watched-dir=/etc/prometheus/rules/prometheus-metric-storage-rulefiles-2],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:reloader-init,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:SHARD,Value:0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:false,MountPath:/etc/prometheus/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-out,ReadOnly:false,MountPath:/etc/prometheus/config_out,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:prometheus-metric-storage-rulefiles-0,ReadOnly:false,MountPath:/etc/prometheus/rules/prometheus-metric-storage-rulefiles-0,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:prometheus-metric-storage-rulefiles-1,ReadOnly:false,MountPath:/etc/prometheus/rules/prometheus-metric-storage-rulefiles-1,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:prometheus-metric-storage-rulefiles-2,ReadOnly:false,MountPath:/etc/prometheus/rules/prometheus-metric-storage-rulefiles-2,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lqpr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod prometheus-metric-storage-0_openstack(5b9177a1-3d3a-4a87-a1fa-1dbe96439288): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 21 08:45:35 crc kubenswrapper[4696]: E0321 08:45:35.667654 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init-config-reloader\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack/prometheus-metric-storage-0" podUID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" Mar 21 08:45:35 crc kubenswrapper[4696]: I0321 08:45:35.912267 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.020832 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4vd4t" event={"ID":"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a","Type":"ContainerStarted","Data":"8579f8303daf9aff53a6e06d9455608fe9f43aa56de937882936fc3bc78e7f6d"} Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.022802 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-ingester-0" event={"ID":"d6518eba-34ec-4a38-b279-cd8bb11e7e24","Type":"ContainerStarted","Data":"f96e6f4eb1a07bff2b08dcf69b191a803a3e0ac7ad81931602da5126488ee2f6"} Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.023734 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" event={"ID":"edf07829-775b-4672-89fe-81c3342300ae","Type":"ContainerStarted","Data":"e562474b89a0fc57b3d419cb25593dc37e66e5aeeb09a5611c28fb1431acf133"} Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.025345 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-index-gateway-0" event={"ID":"88865f3d-9bb9-4844-bb96-dc25c11615ec","Type":"ContainerStarted","Data":"33e0a2ccfb28f9b91bb0ddf3d38615fa60f1fa1869c91acfe96aade59f7a0a6d"} Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.026495 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"dd0d5d6d-801d-4446-a325-f82b2d55bca6","Type":"ContainerStarted","Data":"e64a724dc6fb177174f96e63619f47fb91796f03dbef313eda25e7679afa5d34"} Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.028107 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-compactor-0" event={"ID":"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8","Type":"ContainerStarted","Data":"114545dc83ec20e038a21c521658598dd0ce4a9e9fcfc5b2b2fe1ea1ca836c47"} Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.028970 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" event={"ID":"70a97941-5dea-48a7-8dbf-b475b4af0b6e","Type":"ContainerStarted","Data":"f5721dea1df32ed8b4439f0967327e11f2e76d12018d59b233a5a5b73ab72a40"} Mar 21 08:45:36 crc kubenswrapper[4696]: E0321 08:45:36.030408 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:86738dca2db56373b411adb51a4dfde274968f8c8fde42dbc2daf4bac3bb8daf\\\"\"" pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" podUID="70a97941-5dea-48a7-8dbf-b475b4af0b6e" Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.031412 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5b9177a1-3d3a-4a87-a1fa-1dbe96439288","Type":"ContainerStarted","Data":"fa2c7c7d410cc660805060a6002d5068a2194c994a8895b165f06b0b23a313d7"} Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.036932 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21","Type":"ContainerStarted","Data":"57d05199a17353a9264cfc293d56cc140c724bd2e71321bde21a1d383433484c"} Mar 21 08:45:36 crc kubenswrapper[4696]: E0321 08:45:36.034441 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init-config-reloader\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:e4412f5688c9725f36d2f566f624d82a1a2a5b957686245fd2defcc39604bdc2\\\"\"" pod="openstack/prometheus-metric-storage-0" podUID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.060358 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c1b37b82-bd6a-4e10-98cd-89fbff6bb554","Type":"ContainerStarted","Data":"4fb701a4fed069904e046a153906aef0ff0e084a70050acd1db25ff37cc27178"} Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.069576 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" event={"ID":"bbf61ae0-6918-44af-be69-ac8a220fcd6b","Type":"ContainerStarted","Data":"3caacc515dffaff816b8822d34c29a776abe0638cdcd187ea92f197fb039f3bf"} Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.095379 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" event={"ID":"4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc","Type":"ContainerStarted","Data":"2c75c72854340d1dfeee2b17a2d63890344b0a01617535c9e175e289c1a3f037"} Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.095521 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.097089 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"57e750dc-d19c-45d4-9905-b0a15a6a642d","Type":"ContainerStarted","Data":"8f141f3cfb87e81b9b3e969187a6d57ee85a22b70403d6c81c15eecd6f903a3f"} Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.099461 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" event={"ID":"a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7","Type":"ContainerStarted","Data":"9632e6a0c1a94f1853859311b097b128528719273b1b515fb7416d8699159f95"} Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.105306 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a","Type":"ContainerStarted","Data":"589d42286ee21cc59f6a9512729069f2eff8a5e35c1b64b7c8364d3eb6a79e09"} Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.106761 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" event={"ID":"d2b0b7e1-1fa0-44ce-bab0-0964460e645b","Type":"ContainerStarted","Data":"e85174d41740fe6580a0b0c138bf1d43cc6eb092ec564512e406313b7259c701"} Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.109923 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"b1808d81-bf6f-4201-972a-507d6b41f33f","Type":"ContainerStarted","Data":"78a5db725aa1abd28b9eda81bbf7e747e902fa1839257696a29952ec6d7f797e"} Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.134235 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" podStartSLOduration=15.599042219 podStartE2EDuration="22.134216754s" podCreationTimestamp="2026-03-21 08:45:14 +0000 UTC" firstStartedPulling="2026-03-21 08:45:27.300845601 +0000 UTC m=+1061.421726314" lastFinishedPulling="2026-03-21 08:45:33.836020136 +0000 UTC m=+1067.956900849" observedRunningTime="2026-03-21 08:45:36.112427222 +0000 UTC m=+1070.233307945" watchObservedRunningTime="2026-03-21 08:45:36.134216754 +0000 UTC m=+1070.255097467" Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.547343 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26009008-54e8-48fc-b1b5-99c0e0c1e544" path="/var/lib/kubelet/pods/26009008-54e8-48fc-b1b5-99c0e0c1e544/volumes" Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.547840 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c31577bd-23ed-4ab0-8cf1-42a40b0cc941" path="/var/lib/kubelet/pods/c31577bd-23ed-4ab0-8cf1-42a40b0cc941/volumes" Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.868780 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 21 08:45:36 crc kubenswrapper[4696]: I0321 08:45:36.957682 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-dmbpm"] Mar 21 08:45:37 crc kubenswrapper[4696]: E0321 08:45:37.121497 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init-config-reloader\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:e4412f5688c9725f36d2f566f624d82a1a2a5b957686245fd2defcc39604bdc2\\\"\"" pod="openstack/prometheus-metric-storage-0" podUID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" Mar 21 08:45:37 crc kubenswrapper[4696]: E0321 08:45:37.121784 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gateway\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:86738dca2db56373b411adb51a4dfde274968f8c8fde42dbc2daf4bac3bb8daf\\\"\"" pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" podUID="70a97941-5dea-48a7-8dbf-b475b4af0b6e" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.282659 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-8nzqr"] Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.304396 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-8nzqr" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.316629 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-8nzqr"] Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.324725 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.423311 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bvmvs"] Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.441345 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63645afc-9440-40b7-ae49-9203c9dcb3f8-combined-ca-bundle\") pod \"ovn-controller-metrics-8nzqr\" (UID: \"63645afc-9440-40b7-ae49-9203c9dcb3f8\") " pod="openstack/ovn-controller-metrics-8nzqr" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.441453 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63645afc-9440-40b7-ae49-9203c9dcb3f8-config\") pod \"ovn-controller-metrics-8nzqr\" (UID: \"63645afc-9440-40b7-ae49-9203c9dcb3f8\") " pod="openstack/ovn-controller-metrics-8nzqr" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.441489 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/63645afc-9440-40b7-ae49-9203c9dcb3f8-ovn-rundir\") pod \"ovn-controller-metrics-8nzqr\" (UID: \"63645afc-9440-40b7-ae49-9203c9dcb3f8\") " pod="openstack/ovn-controller-metrics-8nzqr" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.441506 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/63645afc-9440-40b7-ae49-9203c9dcb3f8-ovs-rundir\") pod \"ovn-controller-metrics-8nzqr\" (UID: \"63645afc-9440-40b7-ae49-9203c9dcb3f8\") " pod="openstack/ovn-controller-metrics-8nzqr" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.441622 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqk7p\" (UniqueName: \"kubernetes.io/projected/63645afc-9440-40b7-ae49-9203c9dcb3f8-kube-api-access-mqk7p\") pod \"ovn-controller-metrics-8nzqr\" (UID: \"63645afc-9440-40b7-ae49-9203c9dcb3f8\") " pod="openstack/ovn-controller-metrics-8nzqr" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.441674 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/63645afc-9440-40b7-ae49-9203c9dcb3f8-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-8nzqr\" (UID: \"63645afc-9440-40b7-ae49-9203c9dcb3f8\") " pod="openstack/ovn-controller-metrics-8nzqr" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.446468 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-2lf4b"] Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.447985 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.451399 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.459283 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-2lf4b"] Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.542949 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc1053da-ccb8-46ec-9fb0-370ad118c22a-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-2lf4b\" (UID: \"fc1053da-ccb8-46ec-9fb0-370ad118c22a\") " pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.543011 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63645afc-9440-40b7-ae49-9203c9dcb3f8-config\") pod \"ovn-controller-metrics-8nzqr\" (UID: \"63645afc-9440-40b7-ae49-9203c9dcb3f8\") " pod="openstack/ovn-controller-metrics-8nzqr" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.543053 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/63645afc-9440-40b7-ae49-9203c9dcb3f8-ovn-rundir\") pod \"ovn-controller-metrics-8nzqr\" (UID: \"63645afc-9440-40b7-ae49-9203c9dcb3f8\") " pod="openstack/ovn-controller-metrics-8nzqr" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.543071 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/63645afc-9440-40b7-ae49-9203c9dcb3f8-ovs-rundir\") pod \"ovn-controller-metrics-8nzqr\" (UID: \"63645afc-9440-40b7-ae49-9203c9dcb3f8\") " pod="openstack/ovn-controller-metrics-8nzqr" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.543099 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqk7p\" (UniqueName: \"kubernetes.io/projected/63645afc-9440-40b7-ae49-9203c9dcb3f8-kube-api-access-mqk7p\") pod \"ovn-controller-metrics-8nzqr\" (UID: \"63645afc-9440-40b7-ae49-9203c9dcb3f8\") " pod="openstack/ovn-controller-metrics-8nzqr" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.543118 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/63645afc-9440-40b7-ae49-9203c9dcb3f8-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-8nzqr\" (UID: \"63645afc-9440-40b7-ae49-9203c9dcb3f8\") " pod="openstack/ovn-controller-metrics-8nzqr" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.543138 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jk8q\" (UniqueName: \"kubernetes.io/projected/fc1053da-ccb8-46ec-9fb0-370ad118c22a-kube-api-access-6jk8q\") pod \"dnsmasq-dns-7f896c8c65-2lf4b\" (UID: \"fc1053da-ccb8-46ec-9fb0-370ad118c22a\") " pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.543177 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63645afc-9440-40b7-ae49-9203c9dcb3f8-combined-ca-bundle\") pod \"ovn-controller-metrics-8nzqr\" (UID: \"63645afc-9440-40b7-ae49-9203c9dcb3f8\") " pod="openstack/ovn-controller-metrics-8nzqr" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.543221 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc1053da-ccb8-46ec-9fb0-370ad118c22a-config\") pod \"dnsmasq-dns-7f896c8c65-2lf4b\" (UID: \"fc1053da-ccb8-46ec-9fb0-370ad118c22a\") " pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.543236 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc1053da-ccb8-46ec-9fb0-370ad118c22a-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-2lf4b\" (UID: \"fc1053da-ccb8-46ec-9fb0-370ad118c22a\") " pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.543938 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63645afc-9440-40b7-ae49-9203c9dcb3f8-config\") pod \"ovn-controller-metrics-8nzqr\" (UID: \"63645afc-9440-40b7-ae49-9203c9dcb3f8\") " pod="openstack/ovn-controller-metrics-8nzqr" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.544182 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/63645afc-9440-40b7-ae49-9203c9dcb3f8-ovn-rundir\") pod \"ovn-controller-metrics-8nzqr\" (UID: \"63645afc-9440-40b7-ae49-9203c9dcb3f8\") " pod="openstack/ovn-controller-metrics-8nzqr" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.544229 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/63645afc-9440-40b7-ae49-9203c9dcb3f8-ovs-rundir\") pod \"ovn-controller-metrics-8nzqr\" (UID: \"63645afc-9440-40b7-ae49-9203c9dcb3f8\") " pod="openstack/ovn-controller-metrics-8nzqr" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.559216 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/63645afc-9440-40b7-ae49-9203c9dcb3f8-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-8nzqr\" (UID: \"63645afc-9440-40b7-ae49-9203c9dcb3f8\") " pod="openstack/ovn-controller-metrics-8nzqr" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.560315 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63645afc-9440-40b7-ae49-9203c9dcb3f8-combined-ca-bundle\") pod \"ovn-controller-metrics-8nzqr\" (UID: \"63645afc-9440-40b7-ae49-9203c9dcb3f8\") " pod="openstack/ovn-controller-metrics-8nzqr" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.565838 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqk7p\" (UniqueName: \"kubernetes.io/projected/63645afc-9440-40b7-ae49-9203c9dcb3f8-kube-api-access-mqk7p\") pod \"ovn-controller-metrics-8nzqr\" (UID: \"63645afc-9440-40b7-ae49-9203c9dcb3f8\") " pod="openstack/ovn-controller-metrics-8nzqr" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.617710 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-27j2f"] Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.645906 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc1053da-ccb8-46ec-9fb0-370ad118c22a-config\") pod \"dnsmasq-dns-7f896c8c65-2lf4b\" (UID: \"fc1053da-ccb8-46ec-9fb0-370ad118c22a\") " pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.645956 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc1053da-ccb8-46ec-9fb0-370ad118c22a-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-2lf4b\" (UID: \"fc1053da-ccb8-46ec-9fb0-370ad118c22a\") " pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.645994 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc1053da-ccb8-46ec-9fb0-370ad118c22a-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-2lf4b\" (UID: \"fc1053da-ccb8-46ec-9fb0-370ad118c22a\") " pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.646076 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jk8q\" (UniqueName: \"kubernetes.io/projected/fc1053da-ccb8-46ec-9fb0-370ad118c22a-kube-api-access-6jk8q\") pod \"dnsmasq-dns-7f896c8c65-2lf4b\" (UID: \"fc1053da-ccb8-46ec-9fb0-370ad118c22a\") " pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.647677 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc1053da-ccb8-46ec-9fb0-370ad118c22a-config\") pod \"dnsmasq-dns-7f896c8c65-2lf4b\" (UID: \"fc1053da-ccb8-46ec-9fb0-370ad118c22a\") " pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.648848 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc1053da-ccb8-46ec-9fb0-370ad118c22a-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-2lf4b\" (UID: \"fc1053da-ccb8-46ec-9fb0-370ad118c22a\") " pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.649252 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc1053da-ccb8-46ec-9fb0-370ad118c22a-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-2lf4b\" (UID: \"fc1053da-ccb8-46ec-9fb0-370ad118c22a\") " pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.655748 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7rz9w"] Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.657839 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.660267 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.669657 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7rz9w"] Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.670528 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jk8q\" (UniqueName: \"kubernetes.io/projected/fc1053da-ccb8-46ec-9fb0-370ad118c22a-kube-api-access-6jk8q\") pod \"dnsmasq-dns-7f896c8c65-2lf4b\" (UID: \"fc1053da-ccb8-46ec-9fb0-370ad118c22a\") " pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.674804 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-8nzqr" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.747995 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-7rz9w\" (UID: \"94dbac48-b235-40a3-b76e-4b42661061b1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.748076 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-config\") pod \"dnsmasq-dns-86db49b7ff-7rz9w\" (UID: \"94dbac48-b235-40a3-b76e-4b42661061b1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.748141 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-7rz9w\" (UID: \"94dbac48-b235-40a3-b76e-4b42661061b1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.748178 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp6kq\" (UniqueName: \"kubernetes.io/projected/94dbac48-b235-40a3-b76e-4b42661061b1-kube-api-access-fp6kq\") pod \"dnsmasq-dns-86db49b7ff-7rz9w\" (UID: \"94dbac48-b235-40a3-b76e-4b42661061b1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.748244 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-7rz9w\" (UID: \"94dbac48-b235-40a3-b76e-4b42661061b1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.823900 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.850428 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-config\") pod \"dnsmasq-dns-86db49b7ff-7rz9w\" (UID: \"94dbac48-b235-40a3-b76e-4b42661061b1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.850550 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-7rz9w\" (UID: \"94dbac48-b235-40a3-b76e-4b42661061b1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.850602 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp6kq\" (UniqueName: \"kubernetes.io/projected/94dbac48-b235-40a3-b76e-4b42661061b1-kube-api-access-fp6kq\") pod \"dnsmasq-dns-86db49b7ff-7rz9w\" (UID: \"94dbac48-b235-40a3-b76e-4b42661061b1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.850659 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-7rz9w\" (UID: \"94dbac48-b235-40a3-b76e-4b42661061b1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.850677 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-7rz9w\" (UID: \"94dbac48-b235-40a3-b76e-4b42661061b1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.851469 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-config\") pod \"dnsmasq-dns-86db49b7ff-7rz9w\" (UID: \"94dbac48-b235-40a3-b76e-4b42661061b1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.851968 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-7rz9w\" (UID: \"94dbac48-b235-40a3-b76e-4b42661061b1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.852042 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-7rz9w\" (UID: \"94dbac48-b235-40a3-b76e-4b42661061b1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.852398 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-7rz9w\" (UID: \"94dbac48-b235-40a3-b76e-4b42661061b1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" Mar 21 08:45:37 crc kubenswrapper[4696]: I0321 08:45:37.873978 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp6kq\" (UniqueName: \"kubernetes.io/projected/94dbac48-b235-40a3-b76e-4b42661061b1-kube-api-access-fp6kq\") pod \"dnsmasq-dns-86db49b7ff-7rz9w\" (UID: \"94dbac48-b235-40a3-b76e-4b42661061b1\") " pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" Mar 21 08:45:38 crc kubenswrapper[4696]: I0321 08:45:38.007011 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" Mar 21 08:45:38 crc kubenswrapper[4696]: I0321 08:45:38.127414 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" podUID="4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc" containerName="dnsmasq-dns" containerID="cri-o://2c75c72854340d1dfeee2b17a2d63890344b0a01617535c9e175e289c1a3f037" gracePeriod=10 Mar 21 08:45:39 crc kubenswrapper[4696]: I0321 08:45:39.139862 4696 generic.go:334] "Generic (PLEG): container finished" podID="4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc" containerID="2c75c72854340d1dfeee2b17a2d63890344b0a01617535c9e175e289c1a3f037" exitCode=0 Mar 21 08:45:39 crc kubenswrapper[4696]: I0321 08:45:39.139922 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" event={"ID":"4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc","Type":"ContainerDied","Data":"2c75c72854340d1dfeee2b17a2d63890344b0a01617535c9e175e289c1a3f037"} Mar 21 08:45:40 crc kubenswrapper[4696]: I0321 08:45:40.229958 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" podUID="4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.108:5353: connect: connection refused" Mar 21 08:45:40 crc kubenswrapper[4696]: W0321 08:45:40.857061 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9782617_ee6a_4ab5_aa9c_22dc382f654e.slice/crio-9a593eb195807076639f4d2e1f9f8a46762b9ba023efeed7e27e3854422c5f81 WatchSource:0}: Error finding container 9a593eb195807076639f4d2e1f9f8a46762b9ba023efeed7e27e3854422c5f81: Status 404 returned error can't find the container with id 9a593eb195807076639f4d2e1f9f8a46762b9ba023efeed7e27e3854422c5f81 Mar 21 08:45:41 crc kubenswrapper[4696]: I0321 08:45:41.156172 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"b9782617-ee6a-4ab5-aa9c-22dc382f654e","Type":"ContainerStarted","Data":"9a593eb195807076639f4d2e1f9f8a46762b9ba023efeed7e27e3854422c5f81"} Mar 21 08:45:42 crc kubenswrapper[4696]: W0321 08:45:42.427811 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f0d9af4_d6e3_4168_b83d_5a482fb51d44.slice/crio-5973b336e6a10975dd2a77453a832e0860924bc5084cbe6ff10ad31506ed482b WatchSource:0}: Error finding container 5973b336e6a10975dd2a77453a832e0860924bc5084cbe6ff10ad31506ed482b: Status 404 returned error can't find the container with id 5973b336e6a10975dd2a77453a832e0860924bc5084cbe6ff10ad31506ed482b Mar 21 08:45:43 crc kubenswrapper[4696]: I0321 08:45:43.929196 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dmbpm" event={"ID":"7f0d9af4-d6e3-4168-b83d-5a482fb51d44","Type":"ContainerStarted","Data":"5973b336e6a10975dd2a77453a832e0860924bc5084cbe6ff10ad31506ed482b"} Mar 21 08:45:47 crc kubenswrapper[4696]: I0321 08:45:47.528908 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" Mar 21 08:45:47 crc kubenswrapper[4696]: I0321 08:45:47.647837 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-882nj\" (UniqueName: \"kubernetes.io/projected/4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc-kube-api-access-882nj\") pod \"4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc\" (UID: \"4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc\") " Mar 21 08:45:47 crc kubenswrapper[4696]: I0321 08:45:47.648053 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc-config\") pod \"4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc\" (UID: \"4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc\") " Mar 21 08:45:47 crc kubenswrapper[4696]: I0321 08:45:47.648131 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc-dns-svc\") pod \"4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc\" (UID: \"4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc\") " Mar 21 08:45:47 crc kubenswrapper[4696]: I0321 08:45:47.665703 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc-kube-api-access-882nj" (OuterVolumeSpecName: "kube-api-access-882nj") pod "4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc" (UID: "4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc"). InnerVolumeSpecName "kube-api-access-882nj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:45:47 crc kubenswrapper[4696]: I0321 08:45:47.684618 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc" (UID: "4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:45:47 crc kubenswrapper[4696]: I0321 08:45:47.685459 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc-config" (OuterVolumeSpecName: "config") pod "4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc" (UID: "4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:45:47 crc kubenswrapper[4696]: I0321 08:45:47.750331 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 08:45:47 crc kubenswrapper[4696]: I0321 08:45:47.750366 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-882nj\" (UniqueName: \"kubernetes.io/projected/4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc-kube-api-access-882nj\") on node \"crc\" DevicePath \"\"" Mar 21 08:45:47 crc kubenswrapper[4696]: I0321 08:45:47.750380 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:45:47 crc kubenswrapper[4696]: I0321 08:45:47.964871 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" event={"ID":"4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc","Type":"ContainerDied","Data":"7ef3c3ff0e08ca5b34710de32edf13933a61a288ca5029bbd1744dba9cdd986b"} Mar 21 08:45:47 crc kubenswrapper[4696]: I0321 08:45:47.965123 4696 scope.go:117] "RemoveContainer" containerID="2c75c72854340d1dfeee2b17a2d63890344b0a01617535c9e175e289c1a3f037" Mar 21 08:45:47 crc kubenswrapper[4696]: I0321 08:45:47.964932 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" Mar 21 08:45:48 crc kubenswrapper[4696]: I0321 08:45:48.020441 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bvmvs"] Mar 21 08:45:48 crc kubenswrapper[4696]: I0321 08:45:48.031561 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bvmvs"] Mar 21 08:45:48 crc kubenswrapper[4696]: I0321 08:45:48.544596 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc" path="/var/lib/kubelet/pods/4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc/volumes" Mar 21 08:45:49 crc kubenswrapper[4696]: E0321 08:45:49.681836 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Mar 21 08:45:49 crc kubenswrapper[4696]: E0321 08:45:49.682038 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vtmcd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(57e750dc-d19c-45d4-9905-b0a15a6a642d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 08:45:49 crc kubenswrapper[4696]: E0321 08:45:49.684107 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="57e750dc-d19c-45d4-9905-b0a15a6a642d" Mar 21 08:45:49 crc kubenswrapper[4696]: E0321 08:45:49.689034 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:40a6ef5d62dd6bcd82f3a965d0e00bb5f500b88724f9bc3b06103f1402543b30" Mar 21 08:45:49 crc kubenswrapper[4696]: E0321 08:45:49.689505 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:loki-compactor,Image:registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:40a6ef5d62dd6bcd82f3a965d0e00bb5f500b88724f9bc3b06103f1402543b30,Command:[],Args:[-target=compactor -config.file=/etc/loki/config/config.yaml -runtime-config.file=/etc/loki/config/runtime-config.yaml -config.expand-env=true],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:3100,Protocol:TCP,HostIP:,},ContainerPort{Name:grpclb,HostPort:0,ContainerPort:9095,Protocol:TCP,HostIP:,},ContainerPort{Name:healthchecks,HostPort:0,ContainerPort:3101,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:AWS_ACCESS_KEY_ID,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:cloudkitty-loki-s3,},Key:access_key_id,Optional:nil,},},},EnvVar{Name:AWS_ACCESS_KEY_SECRET,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:cloudkitty-loki-s3,},Key:access_key_secret,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:false,MountPath:/etc/loki/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:storage,ReadOnly:false,MountPath:/tmp/loki,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-compactor-http,ReadOnly:false,MountPath:/var/run/tls/http/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-loki-s3,ReadOnly:false,MountPath:/etc/storage/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-compactor-grpc,ReadOnly:false,MountPath:/var/run/tls/grpc/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-ca-bundle,ReadOnly:false,MountPath:/var/run/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lktqs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/loki/api/v1/status/buildinfo,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:2,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/ready,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-lokistack-compactor-0_openstack(14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 21 08:45:49 crc kubenswrapper[4696]: E0321 08:45:49.690840 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-compactor\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/cloudkitty-lokistack-compactor-0" podUID="14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8" Mar 21 08:45:49 crc kubenswrapper[4696]: E0321 08:45:49.708033 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Mar 21 08:45:49 crc kubenswrapper[4696]: E0321 08:45:49.708253 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v6gkf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(919a11c7-cf69-4660-8de6-6b1a54b63f26): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 08:45:49 crc kubenswrapper[4696]: E0321 08:45:49.710005 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="919a11c7-cf69-4660-8de6-6b1a54b63f26" Mar 21 08:45:49 crc kubenswrapper[4696]: E0321 08:45:49.956693 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified" Mar 21 08:45:49 crc kubenswrapper[4696]: E0321 08:45:49.957093 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovn-controller,Image:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,Command:[ovn-controller --pidfile unix:/run/openvswitch/db.sock --certificate=/etc/pki/tls/certs/ovndb.crt --private-key=/etc/pki/tls/private/ovndb.key --ca-cert=/etc/pki/tls/certs/ovndbca.crt],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5dh657h79h67fh79h694h9dh696hb4h5ffh5c5h5dch56h86h6h557hddhb7h684h5f8h689hc6h5b9hd8h9fh554h666hf8hfch67dh68dh66dq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:var-run,ReadOnly:false,MountPath:/var/run/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-run-ovn,ReadOnly:false,MountPath:/var/run/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-log-ovn,ReadOnly:false,MountPath:/var/log/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cfdl7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_liveness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_readiness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/share/ovn/scripts/ovn-ctl stop_controller],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN SYS_ADMIN SYS_NICE],Drop:[],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-controller-4vd4t_openstack(79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 08:45:49 crc kubenswrapper[4696]: E0321 08:45:49.958441 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovn-controller-4vd4t" podUID="79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a" Mar 21 08:45:49 crc kubenswrapper[4696]: E0321 08:45:49.961378 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:40a6ef5d62dd6bcd82f3a965d0e00bb5f500b88724f9bc3b06103f1402543b30" Mar 21 08:45:49 crc kubenswrapper[4696]: E0321 08:45:49.961496 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:loki-query-frontend,Image:registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:40a6ef5d62dd6bcd82f3a965d0e00bb5f500b88724f9bc3b06103f1402543b30,Command:[],Args:[-target=query-frontend -config.file=/etc/loki/config/config.yaml -runtime-config.file=/etc/loki/config/runtime-config.yaml -config.expand-env=true],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:3100,Protocol:TCP,HostIP:,},ContainerPort{Name:grpclb,HostPort:0,ContainerPort:9095,Protocol:TCP,HostIP:,},ContainerPort{Name:healthchecks,HostPort:0,ContainerPort:3101,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:false,MountPath:/etc/loki/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-query-frontend-http,ReadOnly:false,MountPath:/var/run/tls/http/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-query-frontend-grpc,ReadOnly:false,MountPath:/var/run/tls/grpc/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-ca-bundle,ReadOnly:false,MountPath:/var/run/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dtg8g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/loki/api/v1/status/buildinfo,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:2,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/loki/api/v1/status/buildinfo,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-lokistack-query-frontend-6f54889599-6mn7k_openstack(bbf61ae0-6918-44af-be69-ac8a220fcd6b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 21 08:45:49 crc kubenswrapper[4696]: E0321 08:45:49.962644 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-query-frontend\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" podUID="bbf61ae0-6918-44af-be69-ac8a220fcd6b" Mar 21 08:45:49 crc kubenswrapper[4696]: E0321 08:45:49.995519 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:40a6ef5d62dd6bcd82f3a965d0e00bb5f500b88724f9bc3b06103f1402543b30" Mar 21 08:45:49 crc kubenswrapper[4696]: E0321 08:45:49.995550 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:40a6ef5d62dd6bcd82f3a965d0e00bb5f500b88724f9bc3b06103f1402543b30" Mar 21 08:45:49 crc kubenswrapper[4696]: E0321 08:45:49.995679 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:loki-distributor,Image:registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:40a6ef5d62dd6bcd82f3a965d0e00bb5f500b88724f9bc3b06103f1402543b30,Command:[],Args:[-target=distributor -config.file=/etc/loki/config/config.yaml -runtime-config.file=/etc/loki/config/runtime-config.yaml -config.expand-env=true],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:3100,Protocol:TCP,HostIP:,},ContainerPort{Name:grpclb,HostPort:0,ContainerPort:9095,Protocol:TCP,HostIP:,},ContainerPort{Name:gossip-ring,HostPort:0,ContainerPort:7946,Protocol:TCP,HostIP:,},ContainerPort{Name:healthchecks,HostPort:0,ContainerPort:3101,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:false,MountPath:/etc/loki/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-distributor-http,ReadOnly:false,MountPath:/var/run/tls/http/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-distributor-grpc,ReadOnly:false,MountPath:/var/run/tls/grpc/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-ca-bundle,ReadOnly:false,MountPath:/var/run/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rs2dk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/loki/api/v1/status/buildinfo,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:2,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/ready,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7_openstack(a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 21 08:45:49 crc kubenswrapper[4696]: E0321 08:45:49.995788 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:loki-index-gateway,Image:registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:40a6ef5d62dd6bcd82f3a965d0e00bb5f500b88724f9bc3b06103f1402543b30,Command:[],Args:[-target=index-gateway -config.file=/etc/loki/config/config.yaml -runtime-config.file=/etc/loki/config/runtime-config.yaml -config.expand-env=true],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:3100,Protocol:TCP,HostIP:,},ContainerPort{Name:grpclb,HostPort:0,ContainerPort:9095,Protocol:TCP,HostIP:,},ContainerPort{Name:healthchecks,HostPort:0,ContainerPort:3101,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:AWS_ACCESS_KEY_ID,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:cloudkitty-loki-s3,},Key:access_key_id,Optional:nil,},},},EnvVar{Name:AWS_ACCESS_KEY_SECRET,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:cloudkitty-loki-s3,},Key:access_key_secret,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:false,MountPath:/etc/loki/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:storage,ReadOnly:false,MountPath:/tmp/loki,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-index-gateway-http,ReadOnly:false,MountPath:/var/run/tls/http/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-loki-s3,ReadOnly:false,MountPath:/etc/storage/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-index-gateway-grpc,ReadOnly:false,MountPath:/var/run/tls/grpc/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-ca-bundle,ReadOnly:false,MountPath:/var/run/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xs4kd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/loki/api/v1/status/buildinfo,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:2,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/ready,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-lokistack-index-gateway-0_openstack(88865f3d-9bb9-4844-bb96-dc25c11615ec): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 21 08:45:49 crc kubenswrapper[4696]: E0321 08:45:49.996892 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-distributor\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" podUID="a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7" Mar 21 08:45:49 crc kubenswrapper[4696]: E0321 08:45:49.996986 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-index-gateway\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/cloudkitty-lokistack-index-gateway-0" podUID="88865f3d-9bb9-4844-bb96-dc25c11615ec" Mar 21 08:45:50 crc kubenswrapper[4696]: E0321 08:45:50.005357 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="57e750dc-d19c-45d4-9905-b0a15a6a642d" Mar 21 08:45:50 crc kubenswrapper[4696]: E0321 08:45:50.005446 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-compactor\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:40a6ef5d62dd6bcd82f3a965d0e00bb5f500b88724f9bc3b06103f1402543b30\\\"\"" pod="openstack/cloudkitty-lokistack-compactor-0" podUID="14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8" Mar 21 08:45:50 crc kubenswrapper[4696]: E0321 08:45:50.007104 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="919a11c7-cf69-4660-8de6-6b1a54b63f26" Mar 21 08:45:50 crc kubenswrapper[4696]: E0321 08:45:50.007134 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified\\\"\"" pod="openstack/ovn-controller-4vd4t" podUID="79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a" Mar 21 08:45:50 crc kubenswrapper[4696]: E0321 08:45:50.015068 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-query-frontend\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:40a6ef5d62dd6bcd82f3a965d0e00bb5f500b88724f9bc3b06103f1402543b30\\\"\"" pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" podUID="bbf61ae0-6918-44af-be69-ac8a220fcd6b" Mar 21 08:45:50 crc kubenswrapper[4696]: I0321 08:45:50.226985 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57d769cc4f-bvmvs" podUID="4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.108:5353: i/o timeout" Mar 21 08:45:50 crc kubenswrapper[4696]: I0321 08:45:50.831152 4696 scope.go:117] "RemoveContainer" containerID="de22920663ad6577ced117df929d32519b87df6661cd629b2f26fe803f1f0cc8" Mar 21 08:45:51 crc kubenswrapper[4696]: I0321 08:45:51.359221 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-8nzqr"] Mar 21 08:45:51 crc kubenswrapper[4696]: I0321 08:45:51.366403 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7rz9w"] Mar 21 08:45:51 crc kubenswrapper[4696]: I0321 08:45:51.502044 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-2lf4b"] Mar 21 08:45:51 crc kubenswrapper[4696]: E0321 08:45:51.524595 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Mar 21 08:45:51 crc kubenswrapper[4696]: E0321 08:45:51.524671 4696 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Mar 21 08:45:51 crc kubenswrapper[4696]: E0321 08:45:51.524994 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5dl2g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(c1b37b82-bd6a-4e10-98cd-89fbff6bb554): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 21 08:45:51 crc kubenswrapper[4696]: E0321 08:45:51.526634 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="c1b37b82-bd6a-4e10-98cd-89fbff6bb554" Mar 21 08:45:51 crc kubenswrapper[4696]: W0321 08:45:51.539095 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63645afc_9440_40b7_ae49_9203c9dcb3f8.slice/crio-3f3bfe620a9ecad634e44c4782c87bce2fe5847db62bfee5ea6e1ce75e0a7743 WatchSource:0}: Error finding container 3f3bfe620a9ecad634e44c4782c87bce2fe5847db62bfee5ea6e1ce75e0a7743: Status 404 returned error can't find the container with id 3f3bfe620a9ecad634e44c4782c87bce2fe5847db62bfee5ea6e1ce75e0a7743 Mar 21 08:45:51 crc kubenswrapper[4696]: W0321 08:45:51.540781 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94dbac48_b235_40a3_b76e_4b42661061b1.slice/crio-1729cbc312ad180d8e32384b9c34cd6f46adea276578c0be64cdb7f0db6c690b WatchSource:0}: Error finding container 1729cbc312ad180d8e32384b9c34cd6f46adea276578c0be64cdb7f0db6c690b: Status 404 returned error can't find the container with id 1729cbc312ad180d8e32384b9c34cd6f46adea276578c0be64cdb7f0db6c690b Mar 21 08:45:51 crc kubenswrapper[4696]: W0321 08:45:51.544088 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc1053da_ccb8_46ec_9fb0_370ad118c22a.slice/crio-020a9b8ec89a59e7ea8a26055b55cf13416dcbb81f01f712c0ceb70d6e272662 WatchSource:0}: Error finding container 020a9b8ec89a59e7ea8a26055b55cf13416dcbb81f01f712c0ceb70d6e272662: Status 404 returned error can't find the container with id 020a9b8ec89a59e7ea8a26055b55cf13416dcbb81f01f712c0ceb70d6e272662 Mar 21 08:45:52 crc kubenswrapper[4696]: I0321 08:45:52.035404 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" event={"ID":"94dbac48-b235-40a3-b76e-4b42661061b1","Type":"ContainerStarted","Data":"1729cbc312ad180d8e32384b9c34cd6f46adea276578c0be64cdb7f0db6c690b"} Mar 21 08:45:52 crc kubenswrapper[4696]: I0321 08:45:52.036992 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-8nzqr" event={"ID":"63645afc-9440-40b7-ae49-9203c9dcb3f8","Type":"ContainerStarted","Data":"3f3bfe620a9ecad634e44c4782c87bce2fe5847db62bfee5ea6e1ce75e0a7743"} Mar 21 08:45:52 crc kubenswrapper[4696]: I0321 08:45:52.038350 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" event={"ID":"fc1053da-ccb8-46ec-9fb0-370ad118c22a","Type":"ContainerStarted","Data":"020a9b8ec89a59e7ea8a26055b55cf13416dcbb81f01f712c0ceb70d6e272662"} Mar 21 08:45:52 crc kubenswrapper[4696]: I0321 08:45:52.040204 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" event={"ID":"065e4107-c490-498d-a0f0-fce4011c38f1","Type":"ContainerStarted","Data":"3d200c0dfbaa3b07e65accea28350812647c92d7a181d86ec1e77b8666fdd45f"} Mar 21 08:45:52 crc kubenswrapper[4696]: I0321 08:45:52.040344 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" podUID="065e4107-c490-498d-a0f0-fce4011c38f1" containerName="dnsmasq-dns" containerID="cri-o://3d200c0dfbaa3b07e65accea28350812647c92d7a181d86ec1e77b8666fdd45f" gracePeriod=10 Mar 21 08:45:52 crc kubenswrapper[4696]: I0321 08:45:52.040381 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" Mar 21 08:45:52 crc kubenswrapper[4696]: E0321 08:45:52.042074 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="c1b37b82-bd6a-4e10-98cd-89fbff6bb554" Mar 21 08:45:52 crc kubenswrapper[4696]: I0321 08:45:52.064794 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" podStartSLOduration=22.587572405 podStartE2EDuration="38.064779675s" podCreationTimestamp="2026-03-21 08:45:14 +0000 UTC" firstStartedPulling="2026-03-21 08:45:18.176983094 +0000 UTC m=+1052.297863807" lastFinishedPulling="2026-03-21 08:45:33.654190334 +0000 UTC m=+1067.775071077" observedRunningTime="2026-03-21 08:45:52.058217145 +0000 UTC m=+1086.179097868" watchObservedRunningTime="2026-03-21 08:45:52.064779675 +0000 UTC m=+1086.185660388" Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.048186 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" event={"ID":"70a97941-5dea-48a7-8dbf-b475b4af0b6e","Type":"ContainerStarted","Data":"683f70a874c6161e414f4820ba011efcca193384f89560fa25f37574c00cf924"} Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.049567 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.054014 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" podUID="70a97941-5dea-48a7-8dbf-b475b4af0b6e" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.124:8081/ready\": dial tcp 10.217.0.124:8081: connect: connection refused" Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.054877 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-index-gateway-0" event={"ID":"88865f3d-9bb9-4844-bb96-dc25c11615ec","Type":"ContainerStarted","Data":"ddc4e33d768828447a7523d08a2205b7cb0e12f99e7a6945e713dbff4e8e160a"} Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.055622 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.058130 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21","Type":"ContainerStarted","Data":"e89b350822e3a47b533d2ee5f1e4fc8dfcaeb62fd4c55f5205bdd7f81cc02891"} Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.060599 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"dd0d5d6d-801d-4446-a325-f82b2d55bca6","Type":"ContainerStarted","Data":"b4a5fd395365c0f22e6a57c2d262bc20021eeef797511f943a30f15e3f2c2d88"} Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.061085 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.063063 4696 generic.go:334] "Generic (PLEG): container finished" podID="94dbac48-b235-40a3-b76e-4b42661061b1" containerID="a52a093caa1860a8abb245e742817d3e93a8587d5c48424c078eb2b27413a6f1" exitCode=0 Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.063128 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" event={"ID":"94dbac48-b235-40a3-b76e-4b42661061b1","Type":"ContainerDied","Data":"a52a093caa1860a8abb245e742817d3e93a8587d5c48424c078eb2b27413a6f1"} Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.078637 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" podStartSLOduration=8.769025898 podStartE2EDuration="24.078617838s" podCreationTimestamp="2026-03-21 08:45:29 +0000 UTC" firstStartedPulling="2026-03-21 08:45:35.624632389 +0000 UTC m=+1069.745513092" lastFinishedPulling="2026-03-21 08:45:50.934224319 +0000 UTC m=+1085.055105032" observedRunningTime="2026-03-21 08:45:53.073398155 +0000 UTC m=+1087.194278858" watchObservedRunningTime="2026-03-21 08:45:53.078617838 +0000 UTC m=+1087.199498551" Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.086210 4696 generic.go:334] "Generic (PLEG): container finished" podID="fc1053da-ccb8-46ec-9fb0-370ad118c22a" containerID="072204e582cca157ffd7ee51287af483e123bae48e96fbf3334b791fed02186c" exitCode=0 Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.086275 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" event={"ID":"fc1053da-ccb8-46ec-9fb0-370ad118c22a","Type":"ContainerDied","Data":"072204e582cca157ffd7ee51287af483e123bae48e96fbf3334b791fed02186c"} Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.089309 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-ingester-0" event={"ID":"d6518eba-34ec-4a38-b279-cd8bb11e7e24","Type":"ContainerStarted","Data":"2b1c0cb0a3b630ea880b53bf8aee26efe658b4667f28dc7a08b68f48292c425a"} Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.089493 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.102637 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=19.555266814 podStartE2EDuration="35.102615532s" podCreationTimestamp="2026-03-21 08:45:18 +0000 UTC" firstStartedPulling="2026-03-21 08:45:35.006903816 +0000 UTC m=+1069.127784529" lastFinishedPulling="2026-03-21 08:45:50.554252534 +0000 UTC m=+1084.675133247" observedRunningTime="2026-03-21 08:45:53.099514896 +0000 UTC m=+1087.220395599" watchObservedRunningTime="2026-03-21 08:45:53.102615532 +0000 UTC m=+1087.223496245" Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.110328 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dmbpm" event={"ID":"7f0d9af4-d6e3-4168-b83d-5a482fb51d44","Type":"ContainerStarted","Data":"fe71362f13ab5ec24f2eef4cac20cdbc15bcbc3f540dbb339301d761f07f5a66"} Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.111804 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" event={"ID":"edf07829-775b-4672-89fe-81c3342300ae","Type":"ContainerStarted","Data":"ed711e5a6d311c8e4ddf556c76a92d93ccc583bbca951e06c26bde19014c677b"} Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.112021 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.113791 4696 generic.go:334] "Generic (PLEG): container finished" podID="065e4107-c490-498d-a0f0-fce4011c38f1" containerID="3d200c0dfbaa3b07e65accea28350812647c92d7a181d86ec1e77b8666fdd45f" exitCode=0 Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.113874 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" event={"ID":"065e4107-c490-498d-a0f0-fce4011c38f1","Type":"ContainerDied","Data":"3d200c0dfbaa3b07e65accea28350812647c92d7a181d86ec1e77b8666fdd45f"} Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.113891 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" event={"ID":"065e4107-c490-498d-a0f0-fce4011c38f1","Type":"ContainerDied","Data":"32e469f0a3f6c7870fde7ab3d4bd0355051fe89b9f5755da50661bd7b0af38fd"} Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.113901 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32e469f0a3f6c7870fde7ab3d4bd0355051fe89b9f5755da50661bd7b0af38fd" Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.122671 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-index-gateway-0" podStartSLOduration=-9223372012.732132 podStartE2EDuration="24.122644425s" podCreationTimestamp="2026-03-21 08:45:29 +0000 UTC" firstStartedPulling="2026-03-21 08:45:35.598925258 +0000 UTC m=+1069.719805971" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:45:53.120142306 +0000 UTC m=+1087.241023049" watchObservedRunningTime="2026-03-21 08:45:53.122644425 +0000 UTC m=+1087.243525138" Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.230716 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-ingester-0" podStartSLOduration=8.703348945 podStartE2EDuration="24.230687439s" podCreationTimestamp="2026-03-21 08:45:29 +0000 UTC" firstStartedPulling="2026-03-21 08:45:35.564372795 +0000 UTC m=+1069.685253508" lastFinishedPulling="2026-03-21 08:45:51.091711289 +0000 UTC m=+1085.212592002" observedRunningTime="2026-03-21 08:45:53.164192493 +0000 UTC m=+1087.285073226" watchObservedRunningTime="2026-03-21 08:45:53.230687439 +0000 UTC m=+1087.351568152" Mar 21 08:45:53 crc kubenswrapper[4696]: I0321 08:45:53.274755 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" podStartSLOduration=8.497537579 podStartE2EDuration="24.274733325s" podCreationTimestamp="2026-03-21 08:45:29 +0000 UTC" firstStartedPulling="2026-03-21 08:45:34.990877424 +0000 UTC m=+1069.111758137" lastFinishedPulling="2026-03-21 08:45:50.76807317 +0000 UTC m=+1084.888953883" observedRunningTime="2026-03-21 08:45:53.226449592 +0000 UTC m=+1087.347330315" watchObservedRunningTime="2026-03-21 08:45:53.274733325 +0000 UTC m=+1087.395614038" Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.124150 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" event={"ID":"a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7","Type":"ContainerStarted","Data":"af5e952e94bd46cd606f318a560b5ba41fa79706db350e40c9f9afb93f053566"} Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.124320 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.125555 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"b9782617-ee6a-4ab5-aa9c-22dc382f654e","Type":"ContainerStarted","Data":"36077c29e81bdb0b9c25befd113e21f801a582d54bb66ee5e3fa8f746acce53d"} Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.126964 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" event={"ID":"d2b0b7e1-1fa0-44ce-bab0-0964460e645b","Type":"ContainerStarted","Data":"c4c48ce15010f4dbf33eb8313e0b2c38d8286b698cdb19825927600883014f2b"} Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.127134 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.130376 4696 generic.go:334] "Generic (PLEG): container finished" podID="7f0d9af4-d6e3-4168-b83d-5a482fb51d44" containerID="fe71362f13ab5ec24f2eef4cac20cdbc15bcbc3f540dbb339301d761f07f5a66" exitCode=0 Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.130416 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dmbpm" event={"ID":"7f0d9af4-d6e3-4168-b83d-5a482fb51d44","Type":"ContainerDied","Data":"fe71362f13ab5ec24f2eef4cac20cdbc15bcbc3f540dbb339301d761f07f5a66"} Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.132052 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b057ad21-8030-49e9-b400-c36d433b9f8c","Type":"ContainerStarted","Data":"7fd27604a7d301ac11fa0074a203113325187e8f947c4308c11343eb67ceaed5"} Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.143892 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.148455 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" podStartSLOduration=-9223372011.706362 podStartE2EDuration="25.148413397s" podCreationTimestamp="2026-03-21 08:45:29 +0000 UTC" firstStartedPulling="2026-03-21 08:45:35.597125289 +0000 UTC m=+1069.718006002" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:45:54.141053454 +0000 UTC m=+1088.261934167" watchObservedRunningTime="2026-03-21 08:45:54.148413397 +0000 UTC m=+1088.269294110" Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.156339 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w" Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.166540 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.222590 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-gateway-6b884dc4b5-lczl8" podStartSLOduration=9.277876192 podStartE2EDuration="25.222571595s" podCreationTimestamp="2026-03-21 08:45:29 +0000 UTC" firstStartedPulling="2026-03-21 08:45:35.599896705 +0000 UTC m=+1069.720777408" lastFinishedPulling="2026-03-21 08:45:51.544592098 +0000 UTC m=+1085.665472811" observedRunningTime="2026-03-21 08:45:54.201921615 +0000 UTC m=+1088.322802328" watchObservedRunningTime="2026-03-21 08:45:54.222571595 +0000 UTC m=+1088.343452308" Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.282277 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnr7n\" (UniqueName: \"kubernetes.io/projected/065e4107-c490-498d-a0f0-fce4011c38f1-kube-api-access-tnr7n\") pod \"065e4107-c490-498d-a0f0-fce4011c38f1\" (UID: \"065e4107-c490-498d-a0f0-fce4011c38f1\") " Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.282363 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/065e4107-c490-498d-a0f0-fce4011c38f1-dns-svc\") pod \"065e4107-c490-498d-a0f0-fce4011c38f1\" (UID: \"065e4107-c490-498d-a0f0-fce4011c38f1\") " Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.282401 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/065e4107-c490-498d-a0f0-fce4011c38f1-config\") pod \"065e4107-c490-498d-a0f0-fce4011c38f1\" (UID: \"065e4107-c490-498d-a0f0-fce4011c38f1\") " Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.567989 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/065e4107-c490-498d-a0f0-fce4011c38f1-kube-api-access-tnr7n" (OuterVolumeSpecName: "kube-api-access-tnr7n") pod "065e4107-c490-498d-a0f0-fce4011c38f1" (UID: "065e4107-c490-498d-a0f0-fce4011c38f1"). InnerVolumeSpecName "kube-api-access-tnr7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.596851 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnr7n\" (UniqueName: \"kubernetes.io/projected/065e4107-c490-498d-a0f0-fce4011c38f1-kube-api-access-tnr7n\") on node \"crc\" DevicePath \"\"" Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.610783 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/065e4107-c490-498d-a0f0-fce4011c38f1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "065e4107-c490-498d-a0f0-fce4011c38f1" (UID: "065e4107-c490-498d-a0f0-fce4011c38f1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.615367 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/065e4107-c490-498d-a0f0-fce4011c38f1-config" (OuterVolumeSpecName: "config") pod "065e4107-c490-498d-a0f0-fce4011c38f1" (UID: "065e4107-c490-498d-a0f0-fce4011c38f1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.699239 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/065e4107-c490-498d-a0f0-fce4011c38f1-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 08:45:54 crc kubenswrapper[4696]: I0321 08:45:54.699268 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/065e4107-c490-498d-a0f0-fce4011c38f1-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:45:55 crc kubenswrapper[4696]: I0321 08:45:55.141697 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"b1808d81-bf6f-4201-972a-507d6b41f33f","Type":"ContainerStarted","Data":"96dac74c642b7a9439bc0281c4c3385ddd7bda6594909112429775e3b1fea832"} Mar 21 08:45:55 crc kubenswrapper[4696]: I0321 08:45:55.144179 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5b9177a1-3d3a-4a87-a1fa-1dbe96439288","Type":"ContainerStarted","Data":"853f5cfbde31fd213428791fc1e517f055f22c0aca0b293c96501d865a534f07"} Mar 21 08:45:55 crc kubenswrapper[4696]: I0321 08:45:55.148072 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" event={"ID":"94dbac48-b235-40a3-b76e-4b42661061b1","Type":"ContainerStarted","Data":"2c73f1a127a18cf4511cc1330be78070001d652bb69fb65eb8c52796b6174711"} Mar 21 08:45:55 crc kubenswrapper[4696]: I0321 08:45:55.148125 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" Mar 21 08:45:55 crc kubenswrapper[4696]: I0321 08:45:55.150510 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" event={"ID":"fc1053da-ccb8-46ec-9fb0-370ad118c22a","Type":"ContainerStarted","Data":"2c548c240c00e8989bdc868ddf789d3eeb5102df54da7312e29301b446b92890"} Mar 21 08:45:55 crc kubenswrapper[4696]: I0321 08:45:55.151023 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" Mar 21 08:45:55 crc kubenswrapper[4696]: I0321 08:45:55.152791 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-27j2f" Mar 21 08:45:55 crc kubenswrapper[4696]: I0321 08:45:55.152798 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a","Type":"ContainerStarted","Data":"e8e800b9e6d78c36e85ad2b994e32f91f76eb2ffdc75250bf0f1b501cb1813a8"} Mar 21 08:45:55 crc kubenswrapper[4696]: I0321 08:45:55.183423 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" podStartSLOduration=18.183406235 podStartE2EDuration="18.183406235s" podCreationTimestamp="2026-03-21 08:45:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:45:55.176658018 +0000 UTC m=+1089.297538731" watchObservedRunningTime="2026-03-21 08:45:55.183406235 +0000 UTC m=+1089.304286948" Mar 21 08:45:55 crc kubenswrapper[4696]: I0321 08:45:55.249520 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" podStartSLOduration=18.24950026 podStartE2EDuration="18.24950026s" podCreationTimestamp="2026-03-21 08:45:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:45:55.240350677 +0000 UTC m=+1089.361231390" watchObservedRunningTime="2026-03-21 08:45:55.24950026 +0000 UTC m=+1089.370380973" Mar 21 08:45:55 crc kubenswrapper[4696]: I0321 08:45:55.266305 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-27j2f"] Mar 21 08:45:55 crc kubenswrapper[4696]: I0321 08:45:55.272925 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-27j2f"] Mar 21 08:45:56 crc kubenswrapper[4696]: I0321 08:45:56.550863 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="065e4107-c490-498d-a0f0-fce4011c38f1" path="/var/lib/kubelet/pods/065e4107-c490-498d-a0f0-fce4011c38f1/volumes" Mar 21 08:45:57 crc kubenswrapper[4696]: I0321 08:45:57.169181 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21","Type":"ContainerStarted","Data":"c9a7357c5fd83954d86a738848d1c09dad9b5afafacee97d4c9a7e87995bbb11"} Mar 21 08:45:57 crc kubenswrapper[4696]: I0321 08:45:57.171089 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"b9782617-ee6a-4ab5-aa9c-22dc382f654e","Type":"ContainerStarted","Data":"5dc7ef72dc6811600d904197ca7a17bbc436936fbceeaecd95f4bb61211c77bb"} Mar 21 08:45:57 crc kubenswrapper[4696]: I0321 08:45:57.172909 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-8nzqr" event={"ID":"63645afc-9440-40b7-ae49-9203c9dcb3f8","Type":"ContainerStarted","Data":"90f21168de56135f9ea4bd5e58d3e14d0e1feb90fd0f29a983cf5dd09b7c2de3"} Mar 21 08:45:57 crc kubenswrapper[4696]: I0321 08:45:57.175106 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dmbpm" event={"ID":"7f0d9af4-d6e3-4168-b83d-5a482fb51d44","Type":"ContainerStarted","Data":"369cbda7fdbbb04841f2a2e72a5a8d9a60bb1ff0c262472a88842611f79790b1"} Mar 21 08:45:57 crc kubenswrapper[4696]: I0321 08:45:57.175169 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dmbpm" event={"ID":"7f0d9af4-d6e3-4168-b83d-5a482fb51d44","Type":"ContainerStarted","Data":"9561eb08ff8f59dfa699cdf5c5f7e30328070bd7500572c57fe63fda4fc1675d"} Mar 21 08:45:57 crc kubenswrapper[4696]: I0321 08:45:57.193293 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=9.871035826 podStartE2EDuration="30.193265288s" podCreationTimestamp="2026-03-21 08:45:27 +0000 UTC" firstStartedPulling="2026-03-21 08:45:35.978170674 +0000 UTC m=+1070.099051387" lastFinishedPulling="2026-03-21 08:45:56.300400136 +0000 UTC m=+1090.421280849" observedRunningTime="2026-03-21 08:45:57.184981569 +0000 UTC m=+1091.305862292" watchObservedRunningTime="2026-03-21 08:45:57.193265288 +0000 UTC m=+1091.314146011" Mar 21 08:45:57 crc kubenswrapper[4696]: I0321 08:45:57.210552 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-dmbpm" podStartSLOduration=23.871013642 podStartE2EDuration="32.210532824s" podCreationTimestamp="2026-03-21 08:45:25 +0000 UTC" firstStartedPulling="2026-03-21 08:45:42.43154199 +0000 UTC m=+1076.552422703" lastFinishedPulling="2026-03-21 08:45:50.771061172 +0000 UTC m=+1084.891941885" observedRunningTime="2026-03-21 08:45:57.205560538 +0000 UTC m=+1091.326441251" watchObservedRunningTime="2026-03-21 08:45:57.210532824 +0000 UTC m=+1091.331413537" Mar 21 08:45:57 crc kubenswrapper[4696]: I0321 08:45:57.232257 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=15.844780193 podStartE2EDuration="31.232237654s" podCreationTimestamp="2026-03-21 08:45:26 +0000 UTC" firstStartedPulling="2026-03-21 08:45:40.861298579 +0000 UTC m=+1074.982179292" lastFinishedPulling="2026-03-21 08:45:56.24875604 +0000 UTC m=+1090.369636753" observedRunningTime="2026-03-21 08:45:57.224785239 +0000 UTC m=+1091.345665962" watchObservedRunningTime="2026-03-21 08:45:57.232237654 +0000 UTC m=+1091.353118367" Mar 21 08:45:58 crc kubenswrapper[4696]: I0321 08:45:58.138761 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:58 crc kubenswrapper[4696]: I0321 08:45:58.139189 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:58 crc kubenswrapper[4696]: I0321 08:45:58.190125 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:58 crc kubenswrapper[4696]: I0321 08:45:58.202247 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:45:58 crc kubenswrapper[4696]: I0321 08:45:58.205501 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:58 crc kubenswrapper[4696]: I0321 08:45:58.231340 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-8nzqr" podStartSLOduration=16.526752617 podStartE2EDuration="21.23130844s" podCreationTimestamp="2026-03-21 08:45:37 +0000 UTC" firstStartedPulling="2026-03-21 08:45:51.541210574 +0000 UTC m=+1085.662091287" lastFinishedPulling="2026-03-21 08:45:56.245766397 +0000 UTC m=+1090.366647110" observedRunningTime="2026-03-21 08:45:57.253117961 +0000 UTC m=+1091.373998674" watchObservedRunningTime="2026-03-21 08:45:58.23130844 +0000 UTC m=+1092.352189153" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.014712 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.035223 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.035267 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.094732 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.236419 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.237289 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.559740 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Mar 21 08:45:59 crc kubenswrapper[4696]: E0321 08:45:59.560231 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="065e4107-c490-498d-a0f0-fce4011c38f1" containerName="dnsmasq-dns" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.560254 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="065e4107-c490-498d-a0f0-fce4011c38f1" containerName="dnsmasq-dns" Mar 21 08:45:59 crc kubenswrapper[4696]: E0321 08:45:59.560280 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc" containerName="dnsmasq-dns" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.560288 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc" containerName="dnsmasq-dns" Mar 21 08:45:59 crc kubenswrapper[4696]: E0321 08:45:59.560315 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc" containerName="init" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.560326 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc" containerName="init" Mar 21 08:45:59 crc kubenswrapper[4696]: E0321 08:45:59.560348 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="065e4107-c490-498d-a0f0-fce4011c38f1" containerName="init" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.560355 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="065e4107-c490-498d-a0f0-fce4011c38f1" containerName="init" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.560559 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ed4fc4e-1ec8-41a2-a8fb-d8fbc37071bc" containerName="dnsmasq-dns" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.560588 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="065e4107-c490-498d-a0f0-fce4011c38f1" containerName="dnsmasq-dns" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.561769 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.563952 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-77lb6" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.584887 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.586035 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.586378 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.603746 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.721724 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkwfj\" (UniqueName: \"kubernetes.io/projected/dc5349db-8b6f-41a8-b3c7-d42920c3bf58-kube-api-access-lkwfj\") pod \"ovn-northd-0\" (UID: \"dc5349db-8b6f-41a8-b3c7-d42920c3bf58\") " pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.721853 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc5349db-8b6f-41a8-b3c7-d42920c3bf58-scripts\") pod \"ovn-northd-0\" (UID: \"dc5349db-8b6f-41a8-b3c7-d42920c3bf58\") " pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.721984 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dc5349db-8b6f-41a8-b3c7-d42920c3bf58-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"dc5349db-8b6f-41a8-b3c7-d42920c3bf58\") " pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.722054 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc5349db-8b6f-41a8-b3c7-d42920c3bf58-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"dc5349db-8b6f-41a8-b3c7-d42920c3bf58\") " pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.722124 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc5349db-8b6f-41a8-b3c7-d42920c3bf58-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"dc5349db-8b6f-41a8-b3c7-d42920c3bf58\") " pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.722236 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc5349db-8b6f-41a8-b3c7-d42920c3bf58-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"dc5349db-8b6f-41a8-b3c7-d42920c3bf58\") " pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.722255 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc5349db-8b6f-41a8-b3c7-d42920c3bf58-config\") pod \"ovn-northd-0\" (UID: \"dc5349db-8b6f-41a8-b3c7-d42920c3bf58\") " pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.824348 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc5349db-8b6f-41a8-b3c7-d42920c3bf58-scripts\") pod \"ovn-northd-0\" (UID: \"dc5349db-8b6f-41a8-b3c7-d42920c3bf58\") " pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.824401 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dc5349db-8b6f-41a8-b3c7-d42920c3bf58-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"dc5349db-8b6f-41a8-b3c7-d42920c3bf58\") " pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.824429 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc5349db-8b6f-41a8-b3c7-d42920c3bf58-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"dc5349db-8b6f-41a8-b3c7-d42920c3bf58\") " pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.824484 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc5349db-8b6f-41a8-b3c7-d42920c3bf58-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"dc5349db-8b6f-41a8-b3c7-d42920c3bf58\") " pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.824616 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc5349db-8b6f-41a8-b3c7-d42920c3bf58-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"dc5349db-8b6f-41a8-b3c7-d42920c3bf58\") " pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.824632 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc5349db-8b6f-41a8-b3c7-d42920c3bf58-config\") pod \"ovn-northd-0\" (UID: \"dc5349db-8b6f-41a8-b3c7-d42920c3bf58\") " pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.825574 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkwfj\" (UniqueName: \"kubernetes.io/projected/dc5349db-8b6f-41a8-b3c7-d42920c3bf58-kube-api-access-lkwfj\") pod \"ovn-northd-0\" (UID: \"dc5349db-8b6f-41a8-b3c7-d42920c3bf58\") " pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.825478 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc5349db-8b6f-41a8-b3c7-d42920c3bf58-scripts\") pod \"ovn-northd-0\" (UID: \"dc5349db-8b6f-41a8-b3c7-d42920c3bf58\") " pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.825659 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc5349db-8b6f-41a8-b3c7-d42920c3bf58-config\") pod \"ovn-northd-0\" (UID: \"dc5349db-8b6f-41a8-b3c7-d42920c3bf58\") " pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.825680 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dc5349db-8b6f-41a8-b3c7-d42920c3bf58-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"dc5349db-8b6f-41a8-b3c7-d42920c3bf58\") " pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.831106 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc5349db-8b6f-41a8-b3c7-d42920c3bf58-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"dc5349db-8b6f-41a8-b3c7-d42920c3bf58\") " pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.831788 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc5349db-8b6f-41a8-b3c7-d42920c3bf58-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"dc5349db-8b6f-41a8-b3c7-d42920c3bf58\") " pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.832465 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc5349db-8b6f-41a8-b3c7-d42920c3bf58-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"dc5349db-8b6f-41a8-b3c7-d42920c3bf58\") " pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.844094 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkwfj\" (UniqueName: \"kubernetes.io/projected/dc5349db-8b6f-41a8-b3c7-d42920c3bf58-kube-api-access-lkwfj\") pod \"ovn-northd-0\" (UID: \"dc5349db-8b6f-41a8-b3c7-d42920c3bf58\") " pod="openstack/ovn-northd-0" Mar 21 08:45:59 crc kubenswrapper[4696]: I0321 08:45:59.938992 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 21 08:46:00 crc kubenswrapper[4696]: I0321 08:46:00.136976 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568046-b8pf4"] Mar 21 08:46:00 crc kubenswrapper[4696]: I0321 08:46:00.138975 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568046-b8pf4" Mar 21 08:46:00 crc kubenswrapper[4696]: I0321 08:46:00.141299 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 08:46:00 crc kubenswrapper[4696]: I0321 08:46:00.141461 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 08:46:00 crc kubenswrapper[4696]: I0321 08:46:00.142033 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 08:46:00 crc kubenswrapper[4696]: I0321 08:46:00.153984 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568046-b8pf4"] Mar 21 08:46:00 crc kubenswrapper[4696]: I0321 08:46:00.207850 4696 generic.go:334] "Generic (PLEG): container finished" podID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" containerID="853f5cfbde31fd213428791fc1e517f055f22c0aca0b293c96501d865a534f07" exitCode=0 Mar 21 08:46:00 crc kubenswrapper[4696]: I0321 08:46:00.207937 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5b9177a1-3d3a-4a87-a1fa-1dbe96439288","Type":"ContainerDied","Data":"853f5cfbde31fd213428791fc1e517f055f22c0aca0b293c96501d865a534f07"} Mar 21 08:46:00 crc kubenswrapper[4696]: I0321 08:46:00.210897 4696 generic.go:334] "Generic (PLEG): container finished" podID="b1808d81-bf6f-4201-972a-507d6b41f33f" containerID="96dac74c642b7a9439bc0281c4c3385ddd7bda6594909112429775e3b1fea832" exitCode=0 Mar 21 08:46:00 crc kubenswrapper[4696]: I0321 08:46:00.211046 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"b1808d81-bf6f-4201-972a-507d6b41f33f","Type":"ContainerDied","Data":"96dac74c642b7a9439bc0281c4c3385ddd7bda6594909112429775e3b1fea832"} Mar 21 08:46:00 crc kubenswrapper[4696]: I0321 08:46:00.234958 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6zr7\" (UniqueName: \"kubernetes.io/projected/579a841d-8410-4fa2-bc5f-14aafeaf17d6-kube-api-access-l6zr7\") pod \"auto-csr-approver-29568046-b8pf4\" (UID: \"579a841d-8410-4fa2-bc5f-14aafeaf17d6\") " pod="openshift-infra/auto-csr-approver-29568046-b8pf4" Mar 21 08:46:00 crc kubenswrapper[4696]: I0321 08:46:00.339448 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6zr7\" (UniqueName: \"kubernetes.io/projected/579a841d-8410-4fa2-bc5f-14aafeaf17d6-kube-api-access-l6zr7\") pod \"auto-csr-approver-29568046-b8pf4\" (UID: \"579a841d-8410-4fa2-bc5f-14aafeaf17d6\") " pod="openshift-infra/auto-csr-approver-29568046-b8pf4" Mar 21 08:46:00 crc kubenswrapper[4696]: I0321 08:46:00.389601 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6zr7\" (UniqueName: \"kubernetes.io/projected/579a841d-8410-4fa2-bc5f-14aafeaf17d6-kube-api-access-l6zr7\") pod \"auto-csr-approver-29568046-b8pf4\" (UID: \"579a841d-8410-4fa2-bc5f-14aafeaf17d6\") " pod="openshift-infra/auto-csr-approver-29568046-b8pf4" Mar 21 08:46:00 crc kubenswrapper[4696]: I0321 08:46:00.425086 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 21 08:46:00 crc kubenswrapper[4696]: W0321 08:46:00.428317 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc5349db_8b6f_41a8_b3c7_d42920c3bf58.slice/crio-04ad6b048ecd39ac29eacdfa2302e4e7e6ac952a097790df786eb674d1e55c6a WatchSource:0}: Error finding container 04ad6b048ecd39ac29eacdfa2302e4e7e6ac952a097790df786eb674d1e55c6a: Status 404 returned error can't find the container with id 04ad6b048ecd39ac29eacdfa2302e4e7e6ac952a097790df786eb674d1e55c6a Mar 21 08:46:00 crc kubenswrapper[4696]: I0321 08:46:00.461384 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568046-b8pf4" Mar 21 08:46:00 crc kubenswrapper[4696]: I0321 08:46:00.880494 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568046-b8pf4"] Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.124226 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-2lf4b"] Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.124700 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" podUID="fc1053da-ccb8-46ec-9fb0-370ad118c22a" containerName="dnsmasq-dns" containerID="cri-o://2c548c240c00e8989bdc868ddf789d3eeb5102df54da7312e29301b446b92890" gracePeriod=10 Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.132989 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.189923 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-48vx9"] Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.191366 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-48vx9" Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.242880 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-48vx9"] Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.249297 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"dc5349db-8b6f-41a8-b3c7-d42920c3bf58","Type":"ContainerStarted","Data":"04ad6b048ecd39ac29eacdfa2302e4e7e6ac952a097790df786eb674d1e55c6a"} Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.258245 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568046-b8pf4" event={"ID":"579a841d-8410-4fa2-bc5f-14aafeaf17d6","Type":"ContainerStarted","Data":"332f58aa296fd9657dcc8b814882285baf003e3cef425254bfdd5b80daec0723"} Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.262703 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-48vx9\" (UID: \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\") " pod="openstack/dnsmasq-dns-698758b865-48vx9" Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.262778 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-48vx9\" (UID: \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\") " pod="openstack/dnsmasq-dns-698758b865-48vx9" Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.262938 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzkcc\" (UniqueName: \"kubernetes.io/projected/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-kube-api-access-zzkcc\") pod \"dnsmasq-dns-698758b865-48vx9\" (UID: \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\") " pod="openstack/dnsmasq-dns-698758b865-48vx9" Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.263007 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-dns-svc\") pod \"dnsmasq-dns-698758b865-48vx9\" (UID: \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\") " pod="openstack/dnsmasq-dns-698758b865-48vx9" Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.263054 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-config\") pod \"dnsmasq-dns-698758b865-48vx9\" (UID: \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\") " pod="openstack/dnsmasq-dns-698758b865-48vx9" Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.365830 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-48vx9\" (UID: \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\") " pod="openstack/dnsmasq-dns-698758b865-48vx9" Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.365909 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-48vx9\" (UID: \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\") " pod="openstack/dnsmasq-dns-698758b865-48vx9" Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.366331 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzkcc\" (UniqueName: \"kubernetes.io/projected/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-kube-api-access-zzkcc\") pod \"dnsmasq-dns-698758b865-48vx9\" (UID: \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\") " pod="openstack/dnsmasq-dns-698758b865-48vx9" Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.366504 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-dns-svc\") pod \"dnsmasq-dns-698758b865-48vx9\" (UID: \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\") " pod="openstack/dnsmasq-dns-698758b865-48vx9" Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.366592 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-config\") pod \"dnsmasq-dns-698758b865-48vx9\" (UID: \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\") " pod="openstack/dnsmasq-dns-698758b865-48vx9" Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.367107 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-48vx9\" (UID: \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\") " pod="openstack/dnsmasq-dns-698758b865-48vx9" Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.367456 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-48vx9\" (UID: \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\") " pod="openstack/dnsmasq-dns-698758b865-48vx9" Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.368036 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-dns-svc\") pod \"dnsmasq-dns-698758b865-48vx9\" (UID: \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\") " pod="openstack/dnsmasq-dns-698758b865-48vx9" Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.373693 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-config\") pod \"dnsmasq-dns-698758b865-48vx9\" (UID: \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\") " pod="openstack/dnsmasq-dns-698758b865-48vx9" Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.388022 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzkcc\" (UniqueName: \"kubernetes.io/projected/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-kube-api-access-zzkcc\") pod \"dnsmasq-dns-698758b865-48vx9\" (UID: \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\") " pod="openstack/dnsmasq-dns-698758b865-48vx9" Mar 21 08:46:01 crc kubenswrapper[4696]: I0321 08:46:01.580593 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-48vx9" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.239295 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.310603 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Mar 21 08:46:02 crc kubenswrapper[4696]: E0321 08:46:02.311652 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc1053da-ccb8-46ec-9fb0-370ad118c22a" containerName="dnsmasq-dns" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.311680 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc1053da-ccb8-46ec-9fb0-370ad118c22a" containerName="dnsmasq-dns" Mar 21 08:46:02 crc kubenswrapper[4696]: E0321 08:46:02.311749 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc1053da-ccb8-46ec-9fb0-370ad118c22a" containerName="init" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.311759 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc1053da-ccb8-46ec-9fb0-370ad118c22a" containerName="init" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.316673 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc1053da-ccb8-46ec-9fb0-370ad118c22a" containerName="dnsmasq-dns" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.320229 4696 generic.go:334] "Generic (PLEG): container finished" podID="fc1053da-ccb8-46ec-9fb0-370ad118c22a" containerID="2c548c240c00e8989bdc868ddf789d3eeb5102df54da7312e29301b446b92890" exitCode=0 Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.320658 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.377338 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.377381 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" event={"ID":"fc1053da-ccb8-46ec-9fb0-370ad118c22a","Type":"ContainerDied","Data":"2c548c240c00e8989bdc868ddf789d3eeb5102df54da7312e29301b446b92890"} Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.377407 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-2lf4b" event={"ID":"fc1053da-ccb8-46ec-9fb0-370ad118c22a","Type":"ContainerDied","Data":"020a9b8ec89a59e7ea8a26055b55cf13416dcbb81f01f712c0ceb70d6e272662"} Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.377416 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"919a11c7-cf69-4660-8de6-6b1a54b63f26","Type":"ContainerStarted","Data":"c4cdedafe983caf339bc28809cec361a2ce5d7f72d9fb2372dbab45dfa847d70"} Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.378496 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.378954 4696 scope.go:117] "RemoveContainer" containerID="2c548c240c00e8989bdc868ddf789d3eeb5102df54da7312e29301b446b92890" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.381915 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.382227 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-92zrq" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.382436 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.383649 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.392683 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jk8q\" (UniqueName: \"kubernetes.io/projected/fc1053da-ccb8-46ec-9fb0-370ad118c22a-kube-api-access-6jk8q\") pod \"fc1053da-ccb8-46ec-9fb0-370ad118c22a\" (UID: \"fc1053da-ccb8-46ec-9fb0-370ad118c22a\") " Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.392750 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc1053da-ccb8-46ec-9fb0-370ad118c22a-config\") pod \"fc1053da-ccb8-46ec-9fb0-370ad118c22a\" (UID: \"fc1053da-ccb8-46ec-9fb0-370ad118c22a\") " Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.393158 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc1053da-ccb8-46ec-9fb0-370ad118c22a-ovsdbserver-sb\") pod \"fc1053da-ccb8-46ec-9fb0-370ad118c22a\" (UID: \"fc1053da-ccb8-46ec-9fb0-370ad118c22a\") " Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.394731 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc1053da-ccb8-46ec-9fb0-370ad118c22a-dns-svc\") pod \"fc1053da-ccb8-46ec-9fb0-370ad118c22a\" (UID: \"fc1053da-ccb8-46ec-9fb0-370ad118c22a\") " Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.401610 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc1053da-ccb8-46ec-9fb0-370ad118c22a-kube-api-access-6jk8q" (OuterVolumeSpecName: "kube-api-access-6jk8q") pod "fc1053da-ccb8-46ec-9fb0-370ad118c22a" (UID: "fc1053da-ccb8-46ec-9fb0-370ad118c22a"). InnerVolumeSpecName "kube-api-access-6jk8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.431433 4696 scope.go:117] "RemoveContainer" containerID="072204e582cca157ffd7ee51287af483e123bae48e96fbf3334b791fed02186c" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.458982 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc1053da-ccb8-46ec-9fb0-370ad118c22a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fc1053da-ccb8-46ec-9fb0-370ad118c22a" (UID: "fc1053da-ccb8-46ec-9fb0-370ad118c22a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.461755 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc1053da-ccb8-46ec-9fb0-370ad118c22a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fc1053da-ccb8-46ec-9fb0-370ad118c22a" (UID: "fc1053da-ccb8-46ec-9fb0-370ad118c22a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.467263 4696 scope.go:117] "RemoveContainer" containerID="2c548c240c00e8989bdc868ddf789d3eeb5102df54da7312e29301b446b92890" Mar 21 08:46:02 crc kubenswrapper[4696]: E0321 08:46:02.474049 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c548c240c00e8989bdc868ddf789d3eeb5102df54da7312e29301b446b92890\": container with ID starting with 2c548c240c00e8989bdc868ddf789d3eeb5102df54da7312e29301b446b92890 not found: ID does not exist" containerID="2c548c240c00e8989bdc868ddf789d3eeb5102df54da7312e29301b446b92890" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.474113 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c548c240c00e8989bdc868ddf789d3eeb5102df54da7312e29301b446b92890"} err="failed to get container status \"2c548c240c00e8989bdc868ddf789d3eeb5102df54da7312e29301b446b92890\": rpc error: code = NotFound desc = could not find container \"2c548c240c00e8989bdc868ddf789d3eeb5102df54da7312e29301b446b92890\": container with ID starting with 2c548c240c00e8989bdc868ddf789d3eeb5102df54da7312e29301b446b92890 not found: ID does not exist" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.474151 4696 scope.go:117] "RemoveContainer" containerID="072204e582cca157ffd7ee51287af483e123bae48e96fbf3334b791fed02186c" Mar 21 08:46:02 crc kubenswrapper[4696]: E0321 08:46:02.479728 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"072204e582cca157ffd7ee51287af483e123bae48e96fbf3334b791fed02186c\": container with ID starting with 072204e582cca157ffd7ee51287af483e123bae48e96fbf3334b791fed02186c not found: ID does not exist" containerID="072204e582cca157ffd7ee51287af483e123bae48e96fbf3334b791fed02186c" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.479780 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"072204e582cca157ffd7ee51287af483e123bae48e96fbf3334b791fed02186c"} err="failed to get container status \"072204e582cca157ffd7ee51287af483e123bae48e96fbf3334b791fed02186c\": rpc error: code = NotFound desc = could not find container \"072204e582cca157ffd7ee51287af483e123bae48e96fbf3334b791fed02186c\": container with ID starting with 072204e582cca157ffd7ee51287af483e123bae48e96fbf3334b791fed02186c not found: ID does not exist" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.488589 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-48vx9"] Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.490239 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc1053da-ccb8-46ec-9fb0-370ad118c22a-config" (OuterVolumeSpecName: "config") pod "fc1053da-ccb8-46ec-9fb0-370ad118c22a" (UID: "fc1053da-ccb8-46ec-9fb0-370ad118c22a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.497505 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czhsb\" (UniqueName: \"kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-kube-api-access-czhsb\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.497688 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-etc-swift\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.497790 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2e0c1816-48e6-4cb9-a8e4-32efea57229b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2e0c1816-48e6-4cb9-a8e4-32efea57229b\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.498288 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.498437 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-lock\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.498501 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-cache\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.498630 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jk8q\" (UniqueName: \"kubernetes.io/projected/fc1053da-ccb8-46ec-9fb0-370ad118c22a-kube-api-access-6jk8q\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.498652 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc1053da-ccb8-46ec-9fb0-370ad118c22a-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.498878 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc1053da-ccb8-46ec-9fb0-370ad118c22a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.500937 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc1053da-ccb8-46ec-9fb0-370ad118c22a-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.603068 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-lock\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.603111 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-cache\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.603183 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czhsb\" (UniqueName: \"kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-kube-api-access-czhsb\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.603201 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-etc-swift\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.603218 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2e0c1816-48e6-4cb9-a8e4-32efea57229b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2e0c1816-48e6-4cb9-a8e4-32efea57229b\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.603401 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.603674 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-cache\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.603720 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-lock\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:02 crc kubenswrapper[4696]: E0321 08:46:02.605063 4696 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 21 08:46:02 crc kubenswrapper[4696]: E0321 08:46:02.605083 4696 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 21 08:46:02 crc kubenswrapper[4696]: E0321 08:46:02.605199 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-etc-swift podName:1088ee04-73e8-4aba-aad4-1cd7d26e20aa nodeName:}" failed. No retries permitted until 2026-03-21 08:46:03.105182518 +0000 UTC m=+1097.226063231 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-etc-swift") pod "swift-storage-0" (UID: "1088ee04-73e8-4aba-aad4-1cd7d26e20aa") : configmap "swift-ring-files" not found Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.609093 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.610194 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.610235 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2e0c1816-48e6-4cb9-a8e4-32efea57229b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2e0c1816-48e6-4cb9-a8e4-32efea57229b\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/6664aa90a0202255eed33bf4370d7fe907ab5feb20f7c5883071b2555bdfdd7c/globalmount\"" pod="openstack/swift-storage-0" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.627072 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czhsb\" (UniqueName: \"kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-kube-api-access-czhsb\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.657881 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-2lf4b"] Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.661437 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2e0c1816-48e6-4cb9-a8e4-32efea57229b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2e0c1816-48e6-4cb9-a8e4-32efea57229b\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.664643 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-2lf4b"] Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.842218 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-jznlt"] Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.843725 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.846644 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.847268 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.847537 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.853773 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-jznlt"] Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.908212 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/efd827ca-ae90-4cca-a77d-72fce76c47aa-swiftconf\") pod \"swift-ring-rebalance-jznlt\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.908252 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/efd827ca-ae90-4cca-a77d-72fce76c47aa-scripts\") pod \"swift-ring-rebalance-jznlt\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.908287 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/efd827ca-ae90-4cca-a77d-72fce76c47aa-ring-data-devices\") pod \"swift-ring-rebalance-jznlt\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.908312 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd9v5\" (UniqueName: \"kubernetes.io/projected/efd827ca-ae90-4cca-a77d-72fce76c47aa-kube-api-access-bd9v5\") pod \"swift-ring-rebalance-jznlt\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.908451 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/efd827ca-ae90-4cca-a77d-72fce76c47aa-etc-swift\") pod \"swift-ring-rebalance-jznlt\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.908547 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/efd827ca-ae90-4cca-a77d-72fce76c47aa-dispersionconf\") pod \"swift-ring-rebalance-jznlt\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:02 crc kubenswrapper[4696]: I0321 08:46:02.908635 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efd827ca-ae90-4cca-a77d-72fce76c47aa-combined-ca-bundle\") pod \"swift-ring-rebalance-jznlt\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:03 crc kubenswrapper[4696]: I0321 08:46:03.009034 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" Mar 21 08:46:03 crc kubenswrapper[4696]: I0321 08:46:03.010282 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/efd827ca-ae90-4cca-a77d-72fce76c47aa-ring-data-devices\") pod \"swift-ring-rebalance-jznlt\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:03 crc kubenswrapper[4696]: I0321 08:46:03.010339 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd9v5\" (UniqueName: \"kubernetes.io/projected/efd827ca-ae90-4cca-a77d-72fce76c47aa-kube-api-access-bd9v5\") pod \"swift-ring-rebalance-jznlt\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:03 crc kubenswrapper[4696]: I0321 08:46:03.010591 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/efd827ca-ae90-4cca-a77d-72fce76c47aa-etc-swift\") pod \"swift-ring-rebalance-jznlt\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:03 crc kubenswrapper[4696]: I0321 08:46:03.010634 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/efd827ca-ae90-4cca-a77d-72fce76c47aa-dispersionconf\") pod \"swift-ring-rebalance-jznlt\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:03 crc kubenswrapper[4696]: I0321 08:46:03.010683 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efd827ca-ae90-4cca-a77d-72fce76c47aa-combined-ca-bundle\") pod \"swift-ring-rebalance-jznlt\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:03 crc kubenswrapper[4696]: I0321 08:46:03.010864 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/efd827ca-ae90-4cca-a77d-72fce76c47aa-swiftconf\") pod \"swift-ring-rebalance-jznlt\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:03 crc kubenswrapper[4696]: I0321 08:46:03.010905 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/efd827ca-ae90-4cca-a77d-72fce76c47aa-scripts\") pod \"swift-ring-rebalance-jznlt\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:03 crc kubenswrapper[4696]: I0321 08:46:03.011248 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/efd827ca-ae90-4cca-a77d-72fce76c47aa-etc-swift\") pod \"swift-ring-rebalance-jznlt\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:03 crc kubenswrapper[4696]: I0321 08:46:03.011382 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/efd827ca-ae90-4cca-a77d-72fce76c47aa-ring-data-devices\") pod \"swift-ring-rebalance-jznlt\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:03 crc kubenswrapper[4696]: I0321 08:46:03.011670 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/efd827ca-ae90-4cca-a77d-72fce76c47aa-scripts\") pod \"swift-ring-rebalance-jznlt\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:03 crc kubenswrapper[4696]: I0321 08:46:03.017028 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/efd827ca-ae90-4cca-a77d-72fce76c47aa-dispersionconf\") pod \"swift-ring-rebalance-jznlt\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:03 crc kubenswrapper[4696]: I0321 08:46:03.018743 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efd827ca-ae90-4cca-a77d-72fce76c47aa-combined-ca-bundle\") pod \"swift-ring-rebalance-jznlt\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:03 crc kubenswrapper[4696]: I0321 08:46:03.030517 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/efd827ca-ae90-4cca-a77d-72fce76c47aa-swiftconf\") pod \"swift-ring-rebalance-jznlt\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:03 crc kubenswrapper[4696]: I0321 08:46:03.039676 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd9v5\" (UniqueName: \"kubernetes.io/projected/efd827ca-ae90-4cca-a77d-72fce76c47aa-kube-api-access-bd9v5\") pod \"swift-ring-rebalance-jznlt\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:03 crc kubenswrapper[4696]: I0321 08:46:03.112742 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-etc-swift\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:03 crc kubenswrapper[4696]: E0321 08:46:03.113170 4696 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 21 08:46:03 crc kubenswrapper[4696]: E0321 08:46:03.113189 4696 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 21 08:46:03 crc kubenswrapper[4696]: E0321 08:46:03.113232 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-etc-swift podName:1088ee04-73e8-4aba-aad4-1cd7d26e20aa nodeName:}" failed. No retries permitted until 2026-03-21 08:46:04.1132144 +0000 UTC m=+1098.234095113 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-etc-swift") pod "swift-storage-0" (UID: "1088ee04-73e8-4aba-aad4-1cd7d26e20aa") : configmap "swift-ring-files" not found Mar 21 08:46:03 crc kubenswrapper[4696]: I0321 08:46:03.224357 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:03 crc kubenswrapper[4696]: I0321 08:46:03.334781 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568046-b8pf4" event={"ID":"579a841d-8410-4fa2-bc5f-14aafeaf17d6","Type":"ContainerStarted","Data":"90a6920f56317148709fbc10994ed7dfcbba2b030c0f613beef00e4bbe2fa4b2"} Mar 21 08:46:03 crc kubenswrapper[4696]: I0321 08:46:03.336473 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-48vx9" event={"ID":"782d4c3b-a1bf-4e02-ba5b-0274f885c26b","Type":"ContainerStarted","Data":"cb2812fb454ed3253abf294282aef582ecbccda536b8f0dcfb017acb0015c06d"} Mar 21 08:46:03 crc kubenswrapper[4696]: I0321 08:46:03.337657 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"dc5349db-8b6f-41a8-b3c7-d42920c3bf58","Type":"ContainerStarted","Data":"7082faefd24407a9a52e0eb49775a1d4b5c7119287c64232dec0e7435cf0acf2"} Mar 21 08:46:04 crc kubenswrapper[4696]: I0321 08:46:04.070932 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-jznlt"] Mar 21 08:46:04 crc kubenswrapper[4696]: I0321 08:46:04.130885 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-etc-swift\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:04 crc kubenswrapper[4696]: E0321 08:46:04.131137 4696 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 21 08:46:04 crc kubenswrapper[4696]: E0321 08:46:04.131160 4696 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 21 08:46:04 crc kubenswrapper[4696]: E0321 08:46:04.131200 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-etc-swift podName:1088ee04-73e8-4aba-aad4-1cd7d26e20aa nodeName:}" failed. No retries permitted until 2026-03-21 08:46:06.131187237 +0000 UTC m=+1100.252067950 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-etc-swift") pod "swift-storage-0" (UID: "1088ee04-73e8-4aba-aad4-1cd7d26e20aa") : configmap "swift-ring-files" not found Mar 21 08:46:04 crc kubenswrapper[4696]: I0321 08:46:04.383627 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"dc5349db-8b6f-41a8-b3c7-d42920c3bf58","Type":"ContainerStarted","Data":"19942c9dfcda283fae703d1e3bee26e7758662b2c568f68a1a7af5a305809b7f"} Mar 21 08:46:04 crc kubenswrapper[4696]: I0321 08:46:04.385535 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Mar 21 08:46:04 crc kubenswrapper[4696]: I0321 08:46:04.389064 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" event={"ID":"bbf61ae0-6918-44af-be69-ac8a220fcd6b","Type":"ContainerStarted","Data":"b016c846eae740d1c49d4fc5c3fc78217bbbe7a48eb14ab6564e8b9a6fe6d7cc"} Mar 21 08:46:04 crc kubenswrapper[4696]: I0321 08:46:04.389254 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" Mar 21 08:46:04 crc kubenswrapper[4696]: I0321 08:46:04.394125 4696 generic.go:334] "Generic (PLEG): container finished" podID="579a841d-8410-4fa2-bc5f-14aafeaf17d6" containerID="90a6920f56317148709fbc10994ed7dfcbba2b030c0f613beef00e4bbe2fa4b2" exitCode=0 Mar 21 08:46:04 crc kubenswrapper[4696]: I0321 08:46:04.394234 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568046-b8pf4" event={"ID":"579a841d-8410-4fa2-bc5f-14aafeaf17d6","Type":"ContainerDied","Data":"90a6920f56317148709fbc10994ed7dfcbba2b030c0f613beef00e4bbe2fa4b2"} Mar 21 08:46:04 crc kubenswrapper[4696]: I0321 08:46:04.408644 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-compactor-0" event={"ID":"14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8","Type":"ContainerStarted","Data":"4eab98c8b2d1eada6f55274db08b79b5fc4bdb95040a9b02a864558666e7993e"} Mar 21 08:46:04 crc kubenswrapper[4696]: I0321 08:46:04.409565 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:46:04 crc kubenswrapper[4696]: I0321 08:46:04.412174 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.728083304 podStartE2EDuration="5.412158788s" podCreationTimestamp="2026-03-21 08:45:59 +0000 UTC" firstStartedPulling="2026-03-21 08:46:00.431163461 +0000 UTC m=+1094.552044174" lastFinishedPulling="2026-03-21 08:46:02.115238945 +0000 UTC m=+1096.236119658" observedRunningTime="2026-03-21 08:46:04.402338307 +0000 UTC m=+1098.523219030" watchObservedRunningTime="2026-03-21 08:46:04.412158788 +0000 UTC m=+1098.533039501" Mar 21 08:46:04 crc kubenswrapper[4696]: I0321 08:46:04.415004 4696 generic.go:334] "Generic (PLEG): container finished" podID="782d4c3b-a1bf-4e02-ba5b-0274f885c26b" containerID="97140fe2213ea9b227900bd9eb9d36aa99203d680f911ce09757b1a87aec226d" exitCode=0 Mar 21 08:46:04 crc kubenswrapper[4696]: I0321 08:46:04.415127 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-48vx9" event={"ID":"782d4c3b-a1bf-4e02-ba5b-0274f885c26b","Type":"ContainerDied","Data":"97140fe2213ea9b227900bd9eb9d36aa99203d680f911ce09757b1a87aec226d"} Mar 21 08:46:04 crc kubenswrapper[4696]: I0321 08:46:04.432437 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" podStartSLOduration=-9223372001.422354 podStartE2EDuration="35.432422887s" podCreationTimestamp="2026-03-21 08:45:29 +0000 UTC" firstStartedPulling="2026-03-21 08:45:35.584989783 +0000 UTC m=+1069.705870496" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:46:04.43140956 +0000 UTC m=+1098.552290273" watchObservedRunningTime="2026-03-21 08:46:04.432422887 +0000 UTC m=+1098.553303600" Mar 21 08:46:04 crc kubenswrapper[4696]: I0321 08:46:04.536578 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 08:46:04 crc kubenswrapper[4696]: I0321 08:46:04.547074 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc1053da-ccb8-46ec-9fb0-370ad118c22a" path="/var/lib/kubelet/pods/fc1053da-ccb8-46ec-9fb0-370ad118c22a/volumes" Mar 21 08:46:04 crc kubenswrapper[4696]: I0321 08:46:04.558525 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-compactor-0" podStartSLOduration=-9223372001.296267 podStartE2EDuration="35.55850997s" podCreationTimestamp="2026-03-21 08:45:29 +0000 UTC" firstStartedPulling="2026-03-21 08:45:35.618696234 +0000 UTC m=+1069.739576947" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:46:04.489534155 +0000 UTC m=+1098.610414878" watchObservedRunningTime="2026-03-21 08:46:04.55850997 +0000 UTC m=+1098.679390683" Mar 21 08:46:05 crc kubenswrapper[4696]: W0321 08:46:05.129135 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefd827ca_ae90_4cca_a77d_72fce76c47aa.slice/crio-765ecd3aba1642ab017abeb5841c26f2813b379b4d6c1e6a2adbd9a955b17f63 WatchSource:0}: Error finding container 765ecd3aba1642ab017abeb5841c26f2813b379b4d6c1e6a2adbd9a955b17f63: Status 404 returned error can't find the container with id 765ecd3aba1642ab017abeb5841c26f2813b379b4d6c1e6a2adbd9a955b17f63 Mar 21 08:46:05 crc kubenswrapper[4696]: I0321 08:46:05.434413 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jznlt" event={"ID":"efd827ca-ae90-4cca-a77d-72fce76c47aa","Type":"ContainerStarted","Data":"765ecd3aba1642ab017abeb5841c26f2813b379b4d6c1e6a2adbd9a955b17f63"} Mar 21 08:46:05 crc kubenswrapper[4696]: I0321 08:46:05.766388 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568046-b8pf4" Mar 21 08:46:05 crc kubenswrapper[4696]: I0321 08:46:05.884430 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6zr7\" (UniqueName: \"kubernetes.io/projected/579a841d-8410-4fa2-bc5f-14aafeaf17d6-kube-api-access-l6zr7\") pod \"579a841d-8410-4fa2-bc5f-14aafeaf17d6\" (UID: \"579a841d-8410-4fa2-bc5f-14aafeaf17d6\") " Mar 21 08:46:05 crc kubenswrapper[4696]: I0321 08:46:05.887445 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/579a841d-8410-4fa2-bc5f-14aafeaf17d6-kube-api-access-l6zr7" (OuterVolumeSpecName: "kube-api-access-l6zr7") pod "579a841d-8410-4fa2-bc5f-14aafeaf17d6" (UID: "579a841d-8410-4fa2-bc5f-14aafeaf17d6"). InnerVolumeSpecName "kube-api-access-l6zr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:05 crc kubenswrapper[4696]: I0321 08:46:05.989385 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6zr7\" (UniqueName: \"kubernetes.io/projected/579a841d-8410-4fa2-bc5f-14aafeaf17d6-kube-api-access-l6zr7\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:06 crc kubenswrapper[4696]: I0321 08:46:06.193307 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-etc-swift\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:06 crc kubenswrapper[4696]: E0321 08:46:06.193604 4696 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 21 08:46:06 crc kubenswrapper[4696]: E0321 08:46:06.193632 4696 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 21 08:46:06 crc kubenswrapper[4696]: E0321 08:46:06.193705 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-etc-swift podName:1088ee04-73e8-4aba-aad4-1cd7d26e20aa nodeName:}" failed. No retries permitted until 2026-03-21 08:46:10.193683155 +0000 UTC m=+1104.314563868 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-etc-swift") pod "swift-storage-0" (UID: "1088ee04-73e8-4aba-aad4-1cd7d26e20aa") : configmap "swift-ring-files" not found Mar 21 08:46:06 crc kubenswrapper[4696]: I0321 08:46:06.445077 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"57e750dc-d19c-45d4-9905-b0a15a6a642d","Type":"ContainerStarted","Data":"28db2a3adfa10f21fecad4df0aa53991a17fb568a30bab1bdbcab0b14714e98e"} Mar 21 08:46:06 crc kubenswrapper[4696]: I0321 08:46:06.446919 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568046-b8pf4" event={"ID":"579a841d-8410-4fa2-bc5f-14aafeaf17d6","Type":"ContainerDied","Data":"332f58aa296fd9657dcc8b814882285baf003e3cef425254bfdd5b80daec0723"} Mar 21 08:46:06 crc kubenswrapper[4696]: I0321 08:46:06.446956 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="332f58aa296fd9657dcc8b814882285baf003e3cef425254bfdd5b80daec0723" Mar 21 08:46:06 crc kubenswrapper[4696]: I0321 08:46:06.446996 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568046-b8pf4" Mar 21 08:46:06 crc kubenswrapper[4696]: I0321 08:46:06.448873 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4vd4t" event={"ID":"79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a","Type":"ContainerStarted","Data":"12247bc566379ffdafb9911bed28a46b8a15ec66a9d2a99ca40158f48c2437a1"} Mar 21 08:46:06 crc kubenswrapper[4696]: I0321 08:46:06.449396 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-4vd4t" Mar 21 08:46:06 crc kubenswrapper[4696]: I0321 08:46:06.451906 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-48vx9" event={"ID":"782d4c3b-a1bf-4e02-ba5b-0274f885c26b","Type":"ContainerStarted","Data":"7e7507b773e14ca3d089aa8075b93fec3216569eeeb079b66f4723631ab2cfc1"} Mar 21 08:46:06 crc kubenswrapper[4696]: I0321 08:46:06.452521 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-48vx9" Mar 21 08:46:06 crc kubenswrapper[4696]: I0321 08:46:06.454637 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c1b37b82-bd6a-4e10-98cd-89fbff6bb554","Type":"ContainerStarted","Data":"0660ce817029e70387eb1fe7ee9ba9f46d06dc0936e55f77f894e07fabf086af"} Mar 21 08:46:06 crc kubenswrapper[4696]: I0321 08:46:06.460089 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"b1808d81-bf6f-4201-972a-507d6b41f33f","Type":"ContainerStarted","Data":"25c62ca25ac5a1a5145160c3e83cae831cc0abf528213983218afedbcbdaa662"} Mar 21 08:46:06 crc kubenswrapper[4696]: I0321 08:46:06.535208 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-48vx9" podStartSLOduration=5.535189247 podStartE2EDuration="5.535189247s" podCreationTimestamp="2026-03-21 08:46:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:46:06.49289841 +0000 UTC m=+1100.613779123" watchObservedRunningTime="2026-03-21 08:46:06.535189247 +0000 UTC m=+1100.656069970" Mar 21 08:46:06 crc kubenswrapper[4696]: I0321 08:46:06.542981 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-4vd4t" podStartSLOduration=11.422899827 podStartE2EDuration="41.542968433s" podCreationTimestamp="2026-03-21 08:45:25 +0000 UTC" firstStartedPulling="2026-03-21 08:45:35.1283372 +0000 UTC m=+1069.249217913" lastFinishedPulling="2026-03-21 08:46:05.248405806 +0000 UTC m=+1099.369286519" observedRunningTime="2026-03-21 08:46:06.51750062 +0000 UTC m=+1100.638381353" watchObservedRunningTime="2026-03-21 08:46:06.542968433 +0000 UTC m=+1100.663849146" Mar 21 08:46:06 crc kubenswrapper[4696]: I0321 08:46:06.830645 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568040-n66dl"] Mar 21 08:46:06 crc kubenswrapper[4696]: I0321 08:46:06.840878 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568040-n66dl"] Mar 21 08:46:07 crc kubenswrapper[4696]: I0321 08:46:07.491829 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=16.394195775 podStartE2EDuration="47.49180099s" podCreationTimestamp="2026-03-21 08:45:20 +0000 UTC" firstStartedPulling="2026-03-21 08:45:35.013061837 +0000 UTC m=+1069.133942550" lastFinishedPulling="2026-03-21 08:46:06.110667052 +0000 UTC m=+1100.231547765" observedRunningTime="2026-03-21 08:46:07.485970908 +0000 UTC m=+1101.606851631" watchObservedRunningTime="2026-03-21 08:46:07.49180099 +0000 UTC m=+1101.612681703" Mar 21 08:46:08 crc kubenswrapper[4696]: I0321 08:46:08.556959 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1d6d6a5-6b27-4b94-b754-2544992c5622" path="/var/lib/kubelet/pods/e1d6d6a5-6b27-4b94-b754-2544992c5622/volumes" Mar 21 08:46:09 crc kubenswrapper[4696]: I0321 08:46:09.489096 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"b1808d81-bf6f-4201-972a-507d6b41f33f","Type":"ContainerStarted","Data":"39ce1b922ae61db9210141637d235874aae35e75516b6b7ed43d033461869db8"} Mar 21 08:46:09 crc kubenswrapper[4696]: I0321 08:46:09.489463 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Mar 21 08:46:09 crc kubenswrapper[4696]: I0321 08:46:09.493348 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Mar 21 08:46:09 crc kubenswrapper[4696]: I0321 08:46:09.512179 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7" Mar 21 08:46:09 crc kubenswrapper[4696]: I0321 08:46:09.515129 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=18.268366378 podStartE2EDuration="48.515113663s" podCreationTimestamp="2026-03-21 08:45:21 +0000 UTC" firstStartedPulling="2026-03-21 08:45:35.000782907 +0000 UTC m=+1069.121663630" lastFinishedPulling="2026-03-21 08:46:05.247530202 +0000 UTC m=+1099.368410915" observedRunningTime="2026-03-21 08:46:09.512421507 +0000 UTC m=+1103.633302230" watchObservedRunningTime="2026-03-21 08:46:09.515113663 +0000 UTC m=+1103.635994376" Mar 21 08:46:09 crc kubenswrapper[4696]: I0321 08:46:09.730964 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-querier-668f98fdd7-tnh77" Mar 21 08:46:10 crc kubenswrapper[4696]: I0321 08:46:10.291549 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-etc-swift\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:10 crc kubenswrapper[4696]: E0321 08:46:10.291777 4696 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 21 08:46:10 crc kubenswrapper[4696]: E0321 08:46:10.291842 4696 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 21 08:46:10 crc kubenswrapper[4696]: E0321 08:46:10.291926 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-etc-swift podName:1088ee04-73e8-4aba-aad4-1cd7d26e20aa nodeName:}" failed. No retries permitted until 2026-03-21 08:46:18.291898088 +0000 UTC m=+1112.412778811 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-etc-swift") pod "swift-storage-0" (UID: "1088ee04-73e8-4aba-aad4-1cd7d26e20aa") : configmap "swift-ring-files" not found Mar 21 08:46:10 crc kubenswrapper[4696]: I0321 08:46:10.498497 4696 generic.go:334] "Generic (PLEG): container finished" podID="919a11c7-cf69-4660-8de6-6b1a54b63f26" containerID="c4cdedafe983caf339bc28809cec361a2ce5d7f72d9fb2372dbab45dfa847d70" exitCode=0 Mar 21 08:46:10 crc kubenswrapper[4696]: I0321 08:46:10.498596 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"919a11c7-cf69-4660-8de6-6b1a54b63f26","Type":"ContainerDied","Data":"c4cdedafe983caf339bc28809cec361a2ce5d7f72d9fb2372dbab45dfa847d70"} Mar 21 08:46:10 crc kubenswrapper[4696]: I0321 08:46:10.659139 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="d6518eba-34ec-4a38-b279-cd8bb11e7e24" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 21 08:46:10 crc kubenswrapper[4696]: I0321 08:46:10.885298 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-index-gateway-0" Mar 21 08:46:11 crc kubenswrapper[4696]: I0321 08:46:11.102347 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 21 08:46:11 crc kubenswrapper[4696]: I0321 08:46:11.106430 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 21 08:46:11 crc kubenswrapper[4696]: I0321 08:46:11.582050 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-48vx9" Mar 21 08:46:11 crc kubenswrapper[4696]: I0321 08:46:11.668755 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7rz9w"] Mar 21 08:46:11 crc kubenswrapper[4696]: I0321 08:46:11.668994 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" podUID="94dbac48-b235-40a3-b76e-4b42661061b1" containerName="dnsmasq-dns" containerID="cri-o://2c73f1a127a18cf4511cc1330be78070001d652bb69fb65eb8c52796b6174711" gracePeriod=10 Mar 21 08:46:12 crc kubenswrapper[4696]: I0321 08:46:12.529491 4696 generic.go:334] "Generic (PLEG): container finished" podID="57e750dc-d19c-45d4-9905-b0a15a6a642d" containerID="28db2a3adfa10f21fecad4df0aa53991a17fb568a30bab1bdbcab0b14714e98e" exitCode=0 Mar 21 08:46:12 crc kubenswrapper[4696]: I0321 08:46:12.529780 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"57e750dc-d19c-45d4-9905-b0a15a6a642d","Type":"ContainerDied","Data":"28db2a3adfa10f21fecad4df0aa53991a17fb568a30bab1bdbcab0b14714e98e"} Mar 21 08:46:12 crc kubenswrapper[4696]: I0321 08:46:12.532623 4696 generic.go:334] "Generic (PLEG): container finished" podID="94dbac48-b235-40a3-b76e-4b42661061b1" containerID="2c73f1a127a18cf4511cc1330be78070001d652bb69fb65eb8c52796b6174711" exitCode=0 Mar 21 08:46:12 crc kubenswrapper[4696]: I0321 08:46:12.532651 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" event={"ID":"94dbac48-b235-40a3-b76e-4b42661061b1","Type":"ContainerDied","Data":"2c73f1a127a18cf4511cc1330be78070001d652bb69fb65eb8c52796b6174711"} Mar 21 08:46:12 crc kubenswrapper[4696]: I0321 08:46:12.879953 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.052293 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-dns-svc\") pod \"94dbac48-b235-40a3-b76e-4b42661061b1\" (UID: \"94dbac48-b235-40a3-b76e-4b42661061b1\") " Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.052357 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-ovsdbserver-nb\") pod \"94dbac48-b235-40a3-b76e-4b42661061b1\" (UID: \"94dbac48-b235-40a3-b76e-4b42661061b1\") " Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.052442 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fp6kq\" (UniqueName: \"kubernetes.io/projected/94dbac48-b235-40a3-b76e-4b42661061b1-kube-api-access-fp6kq\") pod \"94dbac48-b235-40a3-b76e-4b42661061b1\" (UID: \"94dbac48-b235-40a3-b76e-4b42661061b1\") " Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.052499 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-ovsdbserver-sb\") pod \"94dbac48-b235-40a3-b76e-4b42661061b1\" (UID: \"94dbac48-b235-40a3-b76e-4b42661061b1\") " Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.052586 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-config\") pod \"94dbac48-b235-40a3-b76e-4b42661061b1\" (UID: \"94dbac48-b235-40a3-b76e-4b42661061b1\") " Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.061070 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94dbac48-b235-40a3-b76e-4b42661061b1-kube-api-access-fp6kq" (OuterVolumeSpecName: "kube-api-access-fp6kq") pod "94dbac48-b235-40a3-b76e-4b42661061b1" (UID: "94dbac48-b235-40a3-b76e-4b42661061b1"). InnerVolumeSpecName "kube-api-access-fp6kq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.100606 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "94dbac48-b235-40a3-b76e-4b42661061b1" (UID: "94dbac48-b235-40a3-b76e-4b42661061b1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.101906 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "94dbac48-b235-40a3-b76e-4b42661061b1" (UID: "94dbac48-b235-40a3-b76e-4b42661061b1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.105598 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-config" (OuterVolumeSpecName: "config") pod "94dbac48-b235-40a3-b76e-4b42661061b1" (UID: "94dbac48-b235-40a3-b76e-4b42661061b1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.109426 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "94dbac48-b235-40a3-b76e-4b42661061b1" (UID: "94dbac48-b235-40a3-b76e-4b42661061b1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.159182 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.159218 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.159231 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fp6kq\" (UniqueName: \"kubernetes.io/projected/94dbac48-b235-40a3-b76e-4b42661061b1-kube-api-access-fp6kq\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.159244 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.159256 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94dbac48-b235-40a3-b76e-4b42661061b1-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.547060 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"919a11c7-cf69-4660-8de6-6b1a54b63f26","Type":"ContainerStarted","Data":"d5b86be846dad5e9ca03be97d761c4fb454fb364bafce383b8fe922230cf2fa8"} Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.548689 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jznlt" event={"ID":"efd827ca-ae90-4cca-a77d-72fce76c47aa","Type":"ContainerStarted","Data":"8c5959bcd8103e1334e9539b249a9e16d933da3b8c91f160f3fd1d0b51a92794"} Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.556342 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5b9177a1-3d3a-4a87-a1fa-1dbe96439288","Type":"ContainerStarted","Data":"6c31373671bc55e8b663291df3d6014ea1784bed879432d8e143689cd7b1d5b8"} Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.559002 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"57e750dc-d19c-45d4-9905-b0a15a6a642d","Type":"ContainerStarted","Data":"d33d875405cf7a5194c1497b5b1eb583d33df219bd96a83b6127f8568f9fd4ac"} Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.563741 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" event={"ID":"94dbac48-b235-40a3-b76e-4b42661061b1","Type":"ContainerDied","Data":"1729cbc312ad180d8e32384b9c34cd6f46adea276578c0be64cdb7f0db6c690b"} Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.563802 4696 scope.go:117] "RemoveContainer" containerID="2c73f1a127a18cf4511cc1330be78070001d652bb69fb65eb8c52796b6174711" Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.563848 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-7rz9w" Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.593242 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=30.842734677 podStartE2EDuration="57.593215153s" podCreationTimestamp="2026-03-21 08:45:16 +0000 UTC" firstStartedPulling="2026-03-21 08:45:34.247527662 +0000 UTC m=+1068.368408375" lastFinishedPulling="2026-03-21 08:46:00.998008138 +0000 UTC m=+1095.118888851" observedRunningTime="2026-03-21 08:46:13.589206801 +0000 UTC m=+1107.710087514" watchObservedRunningTime="2026-03-21 08:46:13.593215153 +0000 UTC m=+1107.714095866" Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.608614 4696 scope.go:117] "RemoveContainer" containerID="a52a093caa1860a8abb245e742817d3e93a8587d5c48424c078eb2b27413a6f1" Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.655134 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-jznlt" podStartSLOduration=4.120635329 podStartE2EDuration="11.65510562s" podCreationTimestamp="2026-03-21 08:46:02 +0000 UTC" firstStartedPulling="2026-03-21 08:46:05.132952247 +0000 UTC m=+1099.253832950" lastFinishedPulling="2026-03-21 08:46:12.667422528 +0000 UTC m=+1106.788303241" observedRunningTime="2026-03-21 08:46:13.618086577 +0000 UTC m=+1107.738967290" watchObservedRunningTime="2026-03-21 08:46:13.65510562 +0000 UTC m=+1107.775986353" Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.655898 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=-9223371980.198893 podStartE2EDuration="56.655883952s" podCreationTimestamp="2026-03-21 08:45:17 +0000 UTC" firstStartedPulling="2026-03-21 08:45:35.062004999 +0000 UTC m=+1069.182885712" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:46:13.642983022 +0000 UTC m=+1107.763863755" watchObservedRunningTime="2026-03-21 08:46:13.655883952 +0000 UTC m=+1107.776764685" Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.684918 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7rz9w"] Mar 21 08:46:13 crc kubenswrapper[4696]: I0321 08:46:13.692442 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7rz9w"] Mar 21 08:46:14 crc kubenswrapper[4696]: I0321 08:46:14.544612 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94dbac48-b235-40a3-b76e-4b42661061b1" path="/var/lib/kubelet/pods/94dbac48-b235-40a3-b76e-4b42661061b1/volumes" Mar 21 08:46:15 crc kubenswrapper[4696]: I0321 08:46:15.588245 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5b9177a1-3d3a-4a87-a1fa-1dbe96439288","Type":"ContainerStarted","Data":"45bca271f037081231d7d8df544b677a84f380931a1ae8f272313aec3e1f6097"} Mar 21 08:46:17 crc kubenswrapper[4696]: I0321 08:46:17.597920 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Mar 21 08:46:17 crc kubenswrapper[4696]: I0321 08:46:17.598216 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 21 08:46:17 crc kubenswrapper[4696]: I0321 08:46:17.680508 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Mar 21 08:46:18 crc kubenswrapper[4696]: I0321 08:46:18.350617 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-etc-swift\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:18 crc kubenswrapper[4696]: E0321 08:46:18.350756 4696 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 21 08:46:18 crc kubenswrapper[4696]: E0321 08:46:18.350841 4696 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 21 08:46:18 crc kubenswrapper[4696]: E0321 08:46:18.350882 4696 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-etc-swift podName:1088ee04-73e8-4aba-aad4-1cd7d26e20aa nodeName:}" failed. No retries permitted until 2026-03-21 08:46:34.350870078 +0000 UTC m=+1128.471750791 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-etc-swift") pod "swift-storage-0" (UID: "1088ee04-73e8-4aba-aad4-1cd7d26e20aa") : configmap "swift-ring-files" not found Mar 21 08:46:18 crc kubenswrapper[4696]: I0321 08:46:18.698516 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Mar 21 08:46:18 crc kubenswrapper[4696]: I0321 08:46:18.817833 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 21 08:46:18 crc kubenswrapper[4696]: I0321 08:46:18.817878 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 21 08:46:18 crc kubenswrapper[4696]: I0321 08:46:18.898535 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.263722 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-4f6e-account-create-update-wx2s5"] Mar 21 08:46:19 crc kubenswrapper[4696]: E0321 08:46:19.264414 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94dbac48-b235-40a3-b76e-4b42661061b1" containerName="init" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.264542 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="94dbac48-b235-40a3-b76e-4b42661061b1" containerName="init" Mar 21 08:46:19 crc kubenswrapper[4696]: E0321 08:46:19.264618 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94dbac48-b235-40a3-b76e-4b42661061b1" containerName="dnsmasq-dns" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.264675 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="94dbac48-b235-40a3-b76e-4b42661061b1" containerName="dnsmasq-dns" Mar 21 08:46:19 crc kubenswrapper[4696]: E0321 08:46:19.264756 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="579a841d-8410-4fa2-bc5f-14aafeaf17d6" containerName="oc" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.264843 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="579a841d-8410-4fa2-bc5f-14aafeaf17d6" containerName="oc" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.265065 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="579a841d-8410-4fa2-bc5f-14aafeaf17d6" containerName="oc" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.265149 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="94dbac48-b235-40a3-b76e-4b42661061b1" containerName="dnsmasq-dns" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.265761 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4f6e-account-create-update-wx2s5" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.268649 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.274809 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-4f6e-account-create-update-wx2s5"] Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.324683 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-nghsz"] Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.325783 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-nghsz" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.332774 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-nghsz"] Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.377610 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d5470b1-6bce-4dac-b1a7-f2ed50bcb713-operator-scripts\") pod \"glance-4f6e-account-create-update-wx2s5\" (UID: \"0d5470b1-6bce-4dac-b1a7-f2ed50bcb713\") " pod="openstack/glance-4f6e-account-create-update-wx2s5" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.377662 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7qd2\" (UniqueName: \"kubernetes.io/projected/0d5470b1-6bce-4dac-b1a7-f2ed50bcb713-kube-api-access-j7qd2\") pod \"glance-4f6e-account-create-update-wx2s5\" (UID: \"0d5470b1-6bce-4dac-b1a7-f2ed50bcb713\") " pod="openstack/glance-4f6e-account-create-update-wx2s5" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.479929 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d5470b1-6bce-4dac-b1a7-f2ed50bcb713-operator-scripts\") pod \"glance-4f6e-account-create-update-wx2s5\" (UID: \"0d5470b1-6bce-4dac-b1a7-f2ed50bcb713\") " pod="openstack/glance-4f6e-account-create-update-wx2s5" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.479985 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7qd2\" (UniqueName: \"kubernetes.io/projected/0d5470b1-6bce-4dac-b1a7-f2ed50bcb713-kube-api-access-j7qd2\") pod \"glance-4f6e-account-create-update-wx2s5\" (UID: \"0d5470b1-6bce-4dac-b1a7-f2ed50bcb713\") " pod="openstack/glance-4f6e-account-create-update-wx2s5" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.480030 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7067c05-68d2-413d-9580-af0eb2d31ef8-operator-scripts\") pod \"glance-db-create-nghsz\" (UID: \"c7067c05-68d2-413d-9580-af0eb2d31ef8\") " pod="openstack/glance-db-create-nghsz" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.480150 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhzk7\" (UniqueName: \"kubernetes.io/projected/c7067c05-68d2-413d-9580-af0eb2d31ef8-kube-api-access-bhzk7\") pod \"glance-db-create-nghsz\" (UID: \"c7067c05-68d2-413d-9580-af0eb2d31ef8\") " pod="openstack/glance-db-create-nghsz" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.481110 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d5470b1-6bce-4dac-b1a7-f2ed50bcb713-operator-scripts\") pod \"glance-4f6e-account-create-update-wx2s5\" (UID: \"0d5470b1-6bce-4dac-b1a7-f2ed50bcb713\") " pod="openstack/glance-4f6e-account-create-update-wx2s5" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.514884 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7qd2\" (UniqueName: \"kubernetes.io/projected/0d5470b1-6bce-4dac-b1a7-f2ed50bcb713-kube-api-access-j7qd2\") pod \"glance-4f6e-account-create-update-wx2s5\" (UID: \"0d5470b1-6bce-4dac-b1a7-f2ed50bcb713\") " pod="openstack/glance-4f6e-account-create-update-wx2s5" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.581893 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7067c05-68d2-413d-9580-af0eb2d31ef8-operator-scripts\") pod \"glance-db-create-nghsz\" (UID: \"c7067c05-68d2-413d-9580-af0eb2d31ef8\") " pod="openstack/glance-db-create-nghsz" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.582007 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhzk7\" (UniqueName: \"kubernetes.io/projected/c7067c05-68d2-413d-9580-af0eb2d31ef8-kube-api-access-bhzk7\") pod \"glance-db-create-nghsz\" (UID: \"c7067c05-68d2-413d-9580-af0eb2d31ef8\") " pod="openstack/glance-db-create-nghsz" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.582461 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4f6e-account-create-update-wx2s5" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.583316 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7067c05-68d2-413d-9580-af0eb2d31ef8-operator-scripts\") pod \"glance-db-create-nghsz\" (UID: \"c7067c05-68d2-413d-9580-af0eb2d31ef8\") " pod="openstack/glance-db-create-nghsz" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.598319 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhzk7\" (UniqueName: \"kubernetes.io/projected/c7067c05-68d2-413d-9580-af0eb2d31ef8-kube-api-access-bhzk7\") pod \"glance-db-create-nghsz\" (UID: \"c7067c05-68d2-413d-9580-af0eb2d31ef8\") " pod="openstack/glance-db-create-nghsz" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.651254 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-nghsz" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.710570 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.861548 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-query-frontend-6f54889599-6mn7k" Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.991799 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-7fs72"] Mar 21 08:46:19 crc kubenswrapper[4696]: I0321 08:46:19.993583 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-7fs72" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.015742 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-7fs72"] Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.016409 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.086920 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-d153-account-create-update-728q9"] Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.091332 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d153-account-create-update-728q9" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.093374 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0936beb9-f71b-4313-b6a0-97b026f697a5-operator-scripts\") pod \"keystone-db-create-7fs72\" (UID: \"0936beb9-f71b-4313-b6a0-97b026f697a5\") " pod="openstack/keystone-db-create-7fs72" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.093429 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj2vq\" (UniqueName: \"kubernetes.io/projected/0936beb9-f71b-4313-b6a0-97b026f697a5-kube-api-access-fj2vq\") pod \"keystone-db-create-7fs72\" (UID: \"0936beb9-f71b-4313-b6a0-97b026f697a5\") " pod="openstack/keystone-db-create-7fs72" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.095395 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.104806 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-d153-account-create-update-728q9"] Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.194726 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0936beb9-f71b-4313-b6a0-97b026f697a5-operator-scripts\") pod \"keystone-db-create-7fs72\" (UID: \"0936beb9-f71b-4313-b6a0-97b026f697a5\") " pod="openstack/keystone-db-create-7fs72" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.194772 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgns8\" (UniqueName: \"kubernetes.io/projected/90f4f716-4da8-4ab0-8757-1c15480c73b5-kube-api-access-fgns8\") pod \"keystone-d153-account-create-update-728q9\" (UID: \"90f4f716-4da8-4ab0-8757-1c15480c73b5\") " pod="openstack/keystone-d153-account-create-update-728q9" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.194829 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj2vq\" (UniqueName: \"kubernetes.io/projected/0936beb9-f71b-4313-b6a0-97b026f697a5-kube-api-access-fj2vq\") pod \"keystone-db-create-7fs72\" (UID: \"0936beb9-f71b-4313-b6a0-97b026f697a5\") " pod="openstack/keystone-db-create-7fs72" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.194862 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90f4f716-4da8-4ab0-8757-1c15480c73b5-operator-scripts\") pod \"keystone-d153-account-create-update-728q9\" (UID: \"90f4f716-4da8-4ab0-8757-1c15480c73b5\") " pod="openstack/keystone-d153-account-create-update-728q9" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.195631 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0936beb9-f71b-4313-b6a0-97b026f697a5-operator-scripts\") pod \"keystone-db-create-7fs72\" (UID: \"0936beb9-f71b-4313-b6a0-97b026f697a5\") " pod="openstack/keystone-db-create-7fs72" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.212293 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj2vq\" (UniqueName: \"kubernetes.io/projected/0936beb9-f71b-4313-b6a0-97b026f697a5-kube-api-access-fj2vq\") pod \"keystone-db-create-7fs72\" (UID: \"0936beb9-f71b-4313-b6a0-97b026f697a5\") " pod="openstack/keystone-db-create-7fs72" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.258295 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-b2hh7"] Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.259638 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-b2hh7" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.269710 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-552e-account-create-update-t4h4f"] Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.270973 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-552e-account-create-update-t4h4f" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.276367 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.278510 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-b2hh7"] Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.296132 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgns8\" (UniqueName: \"kubernetes.io/projected/90f4f716-4da8-4ab0-8757-1c15480c73b5-kube-api-access-fgns8\") pod \"keystone-d153-account-create-update-728q9\" (UID: \"90f4f716-4da8-4ab0-8757-1c15480c73b5\") " pod="openstack/keystone-d153-account-create-update-728q9" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.296512 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90f4f716-4da8-4ab0-8757-1c15480c73b5-operator-scripts\") pod \"keystone-d153-account-create-update-728q9\" (UID: \"90f4f716-4da8-4ab0-8757-1c15480c73b5\") " pod="openstack/keystone-d153-account-create-update-728q9" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.297082 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90f4f716-4da8-4ab0-8757-1c15480c73b5-operator-scripts\") pod \"keystone-d153-account-create-update-728q9\" (UID: \"90f4f716-4da8-4ab0-8757-1c15480c73b5\") " pod="openstack/keystone-d153-account-create-update-728q9" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.298229 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-552e-account-create-update-t4h4f"] Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.327344 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgns8\" (UniqueName: \"kubernetes.io/projected/90f4f716-4da8-4ab0-8757-1c15480c73b5-kube-api-access-fgns8\") pod \"keystone-d153-account-create-update-728q9\" (UID: \"90f4f716-4da8-4ab0-8757-1c15480c73b5\") " pod="openstack/keystone-d153-account-create-update-728q9" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.362101 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-7fs72" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.398719 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7jn4\" (UniqueName: \"kubernetes.io/projected/3c569fb8-edf7-4012-bd27-3eb30c568489-kube-api-access-j7jn4\") pod \"placement-db-create-b2hh7\" (UID: \"3c569fb8-edf7-4012-bd27-3eb30c568489\") " pod="openstack/placement-db-create-b2hh7" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.398773 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0227f04c-42f9-4a76-bd79-c1702200ca30-operator-scripts\") pod \"placement-552e-account-create-update-t4h4f\" (UID: \"0227f04c-42f9-4a76-bd79-c1702200ca30\") " pod="openstack/placement-552e-account-create-update-t4h4f" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.398806 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c569fb8-edf7-4012-bd27-3eb30c568489-operator-scripts\") pod \"placement-db-create-b2hh7\" (UID: \"3c569fb8-edf7-4012-bd27-3eb30c568489\") " pod="openstack/placement-db-create-b2hh7" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.398902 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5ch4\" (UniqueName: \"kubernetes.io/projected/0227f04c-42f9-4a76-bd79-c1702200ca30-kube-api-access-z5ch4\") pod \"placement-552e-account-create-update-t4h4f\" (UID: \"0227f04c-42f9-4a76-bd79-c1702200ca30\") " pod="openstack/placement-552e-account-create-update-t4h4f" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.466010 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d153-account-create-update-728q9" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.500866 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7jn4\" (UniqueName: \"kubernetes.io/projected/3c569fb8-edf7-4012-bd27-3eb30c568489-kube-api-access-j7jn4\") pod \"placement-db-create-b2hh7\" (UID: \"3c569fb8-edf7-4012-bd27-3eb30c568489\") " pod="openstack/placement-db-create-b2hh7" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.500937 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0227f04c-42f9-4a76-bd79-c1702200ca30-operator-scripts\") pod \"placement-552e-account-create-update-t4h4f\" (UID: \"0227f04c-42f9-4a76-bd79-c1702200ca30\") " pod="openstack/placement-552e-account-create-update-t4h4f" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.500976 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c569fb8-edf7-4012-bd27-3eb30c568489-operator-scripts\") pod \"placement-db-create-b2hh7\" (UID: \"3c569fb8-edf7-4012-bd27-3eb30c568489\") " pod="openstack/placement-db-create-b2hh7" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.501022 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5ch4\" (UniqueName: \"kubernetes.io/projected/0227f04c-42f9-4a76-bd79-c1702200ca30-kube-api-access-z5ch4\") pod \"placement-552e-account-create-update-t4h4f\" (UID: \"0227f04c-42f9-4a76-bd79-c1702200ca30\") " pod="openstack/placement-552e-account-create-update-t4h4f" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.501630 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0227f04c-42f9-4a76-bd79-c1702200ca30-operator-scripts\") pod \"placement-552e-account-create-update-t4h4f\" (UID: \"0227f04c-42f9-4a76-bd79-c1702200ca30\") " pod="openstack/placement-552e-account-create-update-t4h4f" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.501860 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c569fb8-edf7-4012-bd27-3eb30c568489-operator-scripts\") pod \"placement-db-create-b2hh7\" (UID: \"3c569fb8-edf7-4012-bd27-3eb30c568489\") " pod="openstack/placement-db-create-b2hh7" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.517033 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5ch4\" (UniqueName: \"kubernetes.io/projected/0227f04c-42f9-4a76-bd79-c1702200ca30-kube-api-access-z5ch4\") pod \"placement-552e-account-create-update-t4h4f\" (UID: \"0227f04c-42f9-4a76-bd79-c1702200ca30\") " pod="openstack/placement-552e-account-create-update-t4h4f" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.520644 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7jn4\" (UniqueName: \"kubernetes.io/projected/3c569fb8-edf7-4012-bd27-3eb30c568489-kube-api-access-j7jn4\") pod \"placement-db-create-b2hh7\" (UID: \"3c569fb8-edf7-4012-bd27-3eb30c568489\") " pod="openstack/placement-db-create-b2hh7" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.583766 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-b2hh7" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.592892 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-552e-account-create-update-t4h4f" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.629686 4696 generic.go:334] "Generic (PLEG): container finished" podID="efd827ca-ae90-4cca-a77d-72fce76c47aa" containerID="8c5959bcd8103e1334e9539b249a9e16d933da3b8c91f160f3fd1d0b51a92794" exitCode=0 Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.629763 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jznlt" event={"ID":"efd827ca-ae90-4cca-a77d-72fce76c47aa","Type":"ContainerDied","Data":"8c5959bcd8103e1334e9539b249a9e16d933da3b8c91f160f3fd1d0b51a92794"} Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.656091 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="d6518eba-34ec-4a38-b279-cd8bb11e7e24" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 21 08:46:20 crc kubenswrapper[4696]: I0321 08:46:20.870195 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-compactor-0" Mar 21 08:46:21 crc kubenswrapper[4696]: I0321 08:46:21.622250 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-nghsz"] Mar 21 08:46:21 crc kubenswrapper[4696]: W0321 08:46:21.622465 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7067c05_68d2_413d_9580_af0eb2d31ef8.slice/crio-21bb64f3b4ae4bbae94f33fad224d256b7f4b1364d493584028ef9d91784db22 WatchSource:0}: Error finding container 21bb64f3b4ae4bbae94f33fad224d256b7f4b1364d493584028ef9d91784db22: Status 404 returned error can't find the container with id 21bb64f3b4ae4bbae94f33fad224d256b7f4b1364d493584028ef9d91784db22 Mar 21 08:46:21 crc kubenswrapper[4696]: I0321 08:46:21.645358 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-nghsz" event={"ID":"c7067c05-68d2-413d-9580-af0eb2d31ef8","Type":"ContainerStarted","Data":"21bb64f3b4ae4bbae94f33fad224d256b7f4b1364d493584028ef9d91784db22"} Mar 21 08:46:21 crc kubenswrapper[4696]: I0321 08:46:21.710294 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-7fs72"] Mar 21 08:46:21 crc kubenswrapper[4696]: I0321 08:46:21.728452 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-d153-account-create-update-728q9"] Mar 21 08:46:21 crc kubenswrapper[4696]: I0321 08:46:21.859805 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-552e-account-create-update-t4h4f"] Mar 21 08:46:21 crc kubenswrapper[4696]: W0321 08:46:21.860059 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d5470b1_6bce_4dac_b1a7_f2ed50bcb713.slice/crio-609243d58967760c92dc69a36f089e3f7e0f43830e4eeee43ccfe1b775dc767e WatchSource:0}: Error finding container 609243d58967760c92dc69a36f089e3f7e0f43830e4eeee43ccfe1b775dc767e: Status 404 returned error can't find the container with id 609243d58967760c92dc69a36f089e3f7e0f43830e4eeee43ccfe1b775dc767e Mar 21 08:46:21 crc kubenswrapper[4696]: I0321 08:46:21.892160 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-4f6e-account-create-update-wx2s5"] Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.032195 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-b2hh7"] Mar 21 08:46:22 crc kubenswrapper[4696]: W0321 08:46:22.032319 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c569fb8_edf7_4012_bd27_3eb30c568489.slice/crio-52905bc118d72d16fff3ae82209e42e1c5c480b0af9606bf6b9424cadb980eeb WatchSource:0}: Error finding container 52905bc118d72d16fff3ae82209e42e1c5c480b0af9606bf6b9424cadb980eeb: Status 404 returned error can't find the container with id 52905bc118d72d16fff3ae82209e42e1c5c480b0af9606bf6b9424cadb980eeb Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.043286 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.144165 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/efd827ca-ae90-4cca-a77d-72fce76c47aa-etc-swift\") pod \"efd827ca-ae90-4cca-a77d-72fce76c47aa\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.144271 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/efd827ca-ae90-4cca-a77d-72fce76c47aa-ring-data-devices\") pod \"efd827ca-ae90-4cca-a77d-72fce76c47aa\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.144293 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/efd827ca-ae90-4cca-a77d-72fce76c47aa-swiftconf\") pod \"efd827ca-ae90-4cca-a77d-72fce76c47aa\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.144344 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/efd827ca-ae90-4cca-a77d-72fce76c47aa-scripts\") pod \"efd827ca-ae90-4cca-a77d-72fce76c47aa\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.144468 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/efd827ca-ae90-4cca-a77d-72fce76c47aa-dispersionconf\") pod \"efd827ca-ae90-4cca-a77d-72fce76c47aa\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.144551 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efd827ca-ae90-4cca-a77d-72fce76c47aa-combined-ca-bundle\") pod \"efd827ca-ae90-4cca-a77d-72fce76c47aa\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.144572 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bd9v5\" (UniqueName: \"kubernetes.io/projected/efd827ca-ae90-4cca-a77d-72fce76c47aa-kube-api-access-bd9v5\") pod \"efd827ca-ae90-4cca-a77d-72fce76c47aa\" (UID: \"efd827ca-ae90-4cca-a77d-72fce76c47aa\") " Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.146338 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efd827ca-ae90-4cca-a77d-72fce76c47aa-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "efd827ca-ae90-4cca-a77d-72fce76c47aa" (UID: "efd827ca-ae90-4cca-a77d-72fce76c47aa"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.146914 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efd827ca-ae90-4cca-a77d-72fce76c47aa-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "efd827ca-ae90-4cca-a77d-72fce76c47aa" (UID: "efd827ca-ae90-4cca-a77d-72fce76c47aa"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.163509 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efd827ca-ae90-4cca-a77d-72fce76c47aa-kube-api-access-bd9v5" (OuterVolumeSpecName: "kube-api-access-bd9v5") pod "efd827ca-ae90-4cca-a77d-72fce76c47aa" (UID: "efd827ca-ae90-4cca-a77d-72fce76c47aa"). InnerVolumeSpecName "kube-api-access-bd9v5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.181809 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efd827ca-ae90-4cca-a77d-72fce76c47aa-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "efd827ca-ae90-4cca-a77d-72fce76c47aa" (UID: "efd827ca-ae90-4cca-a77d-72fce76c47aa"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.248190 4696 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/efd827ca-ae90-4cca-a77d-72fce76c47aa-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.248233 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bd9v5\" (UniqueName: \"kubernetes.io/projected/efd827ca-ae90-4cca-a77d-72fce76c47aa-kube-api-access-bd9v5\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.248247 4696 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/efd827ca-ae90-4cca-a77d-72fce76c47aa-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.248258 4696 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/efd827ca-ae90-4cca-a77d-72fce76c47aa-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.250939 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efd827ca-ae90-4cca-a77d-72fce76c47aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "efd827ca-ae90-4cca-a77d-72fce76c47aa" (UID: "efd827ca-ae90-4cca-a77d-72fce76c47aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.259026 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efd827ca-ae90-4cca-a77d-72fce76c47aa-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "efd827ca-ae90-4cca-a77d-72fce76c47aa" (UID: "efd827ca-ae90-4cca-a77d-72fce76c47aa"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.280770 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efd827ca-ae90-4cca-a77d-72fce76c47aa-scripts" (OuterVolumeSpecName: "scripts") pod "efd827ca-ae90-4cca-a77d-72fce76c47aa" (UID: "efd827ca-ae90-4cca-a77d-72fce76c47aa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.349643 4696 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/efd827ca-ae90-4cca-a77d-72fce76c47aa-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.349671 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/efd827ca-ae90-4cca-a77d-72fce76c47aa-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.349681 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efd827ca-ae90-4cca-a77d-72fce76c47aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.653592 4696 generic.go:334] "Generic (PLEG): container finished" podID="0227f04c-42f9-4a76-bd79-c1702200ca30" containerID="477c1fef1a311a6dee8cbdd3fb10286d5c2527cdd12e07df175ced7c294a0f79" exitCode=0 Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.653776 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-552e-account-create-update-t4h4f" event={"ID":"0227f04c-42f9-4a76-bd79-c1702200ca30","Type":"ContainerDied","Data":"477c1fef1a311a6dee8cbdd3fb10286d5c2527cdd12e07df175ced7c294a0f79"} Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.654617 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-552e-account-create-update-t4h4f" event={"ID":"0227f04c-42f9-4a76-bd79-c1702200ca30","Type":"ContainerStarted","Data":"c47951d0a611160aeb80af4ed58e6a20060a5db2360b2e2227b3f47a16aa51cd"} Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.660030 4696 generic.go:334] "Generic (PLEG): container finished" podID="90f4f716-4da8-4ab0-8757-1c15480c73b5" containerID="165c8871ae76943c82c5256046d6b9992394e044b7f56566c6507b0d68eff7ac" exitCode=0 Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.660278 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d153-account-create-update-728q9" event={"ID":"90f4f716-4da8-4ab0-8757-1c15480c73b5","Type":"ContainerDied","Data":"165c8871ae76943c82c5256046d6b9992394e044b7f56566c6507b0d68eff7ac"} Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.660325 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d153-account-create-update-728q9" event={"ID":"90f4f716-4da8-4ab0-8757-1c15480c73b5","Type":"ContainerStarted","Data":"0e08943a2c58e15cdc71e51cf5c443f34068fa535cf3e741dcaa9978f5150e2d"} Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.663841 4696 generic.go:334] "Generic (PLEG): container finished" podID="0d5470b1-6bce-4dac-b1a7-f2ed50bcb713" containerID="c2dd58cbddac3e8aa6f911a814016b9eb28687b142952a91d917b2d3bf923f6d" exitCode=0 Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.664006 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-4f6e-account-create-update-wx2s5" event={"ID":"0d5470b1-6bce-4dac-b1a7-f2ed50bcb713","Type":"ContainerDied","Data":"c2dd58cbddac3e8aa6f911a814016b9eb28687b142952a91d917b2d3bf923f6d"} Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.664032 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-4f6e-account-create-update-wx2s5" event={"ID":"0d5470b1-6bce-4dac-b1a7-f2ed50bcb713","Type":"ContainerStarted","Data":"609243d58967760c92dc69a36f089e3f7e0f43830e4eeee43ccfe1b775dc767e"} Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.673282 4696 generic.go:334] "Generic (PLEG): container finished" podID="0936beb9-f71b-4313-b6a0-97b026f697a5" containerID="090b80904642dc7de45b60aa7b56752e6360346514061cda1774d4693d550e96" exitCode=0 Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.673391 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-7fs72" event={"ID":"0936beb9-f71b-4313-b6a0-97b026f697a5","Type":"ContainerDied","Data":"090b80904642dc7de45b60aa7b56752e6360346514061cda1774d4693d550e96"} Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.673498 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-7fs72" event={"ID":"0936beb9-f71b-4313-b6a0-97b026f697a5","Type":"ContainerStarted","Data":"2f929285f063c9e68a989915acb4cd7824c5e44d04924c3682fe8299dcd1f701"} Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.676648 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jznlt" Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.676949 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jznlt" event={"ID":"efd827ca-ae90-4cca-a77d-72fce76c47aa","Type":"ContainerDied","Data":"765ecd3aba1642ab017abeb5841c26f2813b379b4d6c1e6a2adbd9a955b17f63"} Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.676983 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="765ecd3aba1642ab017abeb5841c26f2813b379b4d6c1e6a2adbd9a955b17f63" Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.680642 4696 generic.go:334] "Generic (PLEG): container finished" podID="3c569fb8-edf7-4012-bd27-3eb30c568489" containerID="bd9b084d944210fce522b4d40b67d3fe59c3becdacac46cf7e660c5b96ede3d1" exitCode=0 Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.680706 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-b2hh7" event={"ID":"3c569fb8-edf7-4012-bd27-3eb30c568489","Type":"ContainerDied","Data":"bd9b084d944210fce522b4d40b67d3fe59c3becdacac46cf7e660c5b96ede3d1"} Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.680735 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-b2hh7" event={"ID":"3c569fb8-edf7-4012-bd27-3eb30c568489","Type":"ContainerStarted","Data":"52905bc118d72d16fff3ae82209e42e1c5c480b0af9606bf6b9424cadb980eeb"} Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.688052 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5b9177a1-3d3a-4a87-a1fa-1dbe96439288","Type":"ContainerStarted","Data":"10d17fe0abf4d917f53465a679a5fbdd91f1967736cd68e4446a6853891177b2"} Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.692071 4696 generic.go:334] "Generic (PLEG): container finished" podID="c7067c05-68d2-413d-9580-af0eb2d31ef8" containerID="440899af6e87a066f77adc8aeb0aab34cd3a0ed754e6b67b3a2c4b827aeee757" exitCode=0 Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.692549 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-nghsz" event={"ID":"c7067c05-68d2-413d-9580-af0eb2d31ef8","Type":"ContainerDied","Data":"440899af6e87a066f77adc8aeb0aab34cd3a0ed754e6b67b3a2c4b827aeee757"} Mar 21 08:46:22 crc kubenswrapper[4696]: I0321 08:46:22.766338 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=15.383750551 podStartE2EDuration="1m1.766322494s" podCreationTimestamp="2026-03-21 08:45:21 +0000 UTC" firstStartedPulling="2026-03-21 08:45:35.666169686 +0000 UTC m=+1069.787050399" lastFinishedPulling="2026-03-21 08:46:22.048741629 +0000 UTC m=+1116.169622342" observedRunningTime="2026-03-21 08:46:22.749302179 +0000 UTC m=+1116.870182882" watchObservedRunningTime="2026-03-21 08:46:22.766322494 +0000 UTC m=+1116.887203207" Mar 21 08:46:23 crc kubenswrapper[4696]: I0321 08:46:23.188449 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:23 crc kubenswrapper[4696]: I0321 08:46:23.188492 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:23 crc kubenswrapper[4696]: I0321 08:46:23.191174 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:23 crc kubenswrapper[4696]: I0321 08:46:23.702531 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.131722 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d153-account-create-update-728q9" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.180132 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgns8\" (UniqueName: \"kubernetes.io/projected/90f4f716-4da8-4ab0-8757-1c15480c73b5-kube-api-access-fgns8\") pod \"90f4f716-4da8-4ab0-8757-1c15480c73b5\" (UID: \"90f4f716-4da8-4ab0-8757-1c15480c73b5\") " Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.180305 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90f4f716-4da8-4ab0-8757-1c15480c73b5-operator-scripts\") pod \"90f4f716-4da8-4ab0-8757-1c15480c73b5\" (UID: \"90f4f716-4da8-4ab0-8757-1c15480c73b5\") " Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.181119 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90f4f716-4da8-4ab0-8757-1c15480c73b5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "90f4f716-4da8-4ab0-8757-1c15480c73b5" (UID: "90f4f716-4da8-4ab0-8757-1c15480c73b5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.204063 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90f4f716-4da8-4ab0-8757-1c15480c73b5-kube-api-access-fgns8" (OuterVolumeSpecName: "kube-api-access-fgns8") pod "90f4f716-4da8-4ab0-8757-1c15480c73b5" (UID: "90f4f716-4da8-4ab0-8757-1c15480c73b5"). InnerVolumeSpecName "kube-api-access-fgns8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.282954 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90f4f716-4da8-4ab0-8757-1c15480c73b5-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.283000 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgns8\" (UniqueName: \"kubernetes.io/projected/90f4f716-4da8-4ab0-8757-1c15480c73b5-kube-api-access-fgns8\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.330660 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-nghsz" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.335115 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4f6e-account-create-update-wx2s5" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.345355 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-b2hh7" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.354804 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-552e-account-create-update-t4h4f" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.364937 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-7fs72" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.383934 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d5470b1-6bce-4dac-b1a7-f2ed50bcb713-operator-scripts\") pod \"0d5470b1-6bce-4dac-b1a7-f2ed50bcb713\" (UID: \"0d5470b1-6bce-4dac-b1a7-f2ed50bcb713\") " Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.384034 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhzk7\" (UniqueName: \"kubernetes.io/projected/c7067c05-68d2-413d-9580-af0eb2d31ef8-kube-api-access-bhzk7\") pod \"c7067c05-68d2-413d-9580-af0eb2d31ef8\" (UID: \"c7067c05-68d2-413d-9580-af0eb2d31ef8\") " Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.384057 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7jn4\" (UniqueName: \"kubernetes.io/projected/3c569fb8-edf7-4012-bd27-3eb30c568489-kube-api-access-j7jn4\") pod \"3c569fb8-edf7-4012-bd27-3eb30c568489\" (UID: \"3c569fb8-edf7-4012-bd27-3eb30c568489\") " Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.384139 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c569fb8-edf7-4012-bd27-3eb30c568489-operator-scripts\") pod \"3c569fb8-edf7-4012-bd27-3eb30c568489\" (UID: \"3c569fb8-edf7-4012-bd27-3eb30c568489\") " Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.384171 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7qd2\" (UniqueName: \"kubernetes.io/projected/0d5470b1-6bce-4dac-b1a7-f2ed50bcb713-kube-api-access-j7qd2\") pod \"0d5470b1-6bce-4dac-b1a7-f2ed50bcb713\" (UID: \"0d5470b1-6bce-4dac-b1a7-f2ed50bcb713\") " Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.384219 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7067c05-68d2-413d-9580-af0eb2d31ef8-operator-scripts\") pod \"c7067c05-68d2-413d-9580-af0eb2d31ef8\" (UID: \"c7067c05-68d2-413d-9580-af0eb2d31ef8\") " Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.384457 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d5470b1-6bce-4dac-b1a7-f2ed50bcb713-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0d5470b1-6bce-4dac-b1a7-f2ed50bcb713" (UID: "0d5470b1-6bce-4dac-b1a7-f2ed50bcb713"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.384788 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d5470b1-6bce-4dac-b1a7-f2ed50bcb713-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.384803 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7067c05-68d2-413d-9580-af0eb2d31ef8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c7067c05-68d2-413d-9580-af0eb2d31ef8" (UID: "c7067c05-68d2-413d-9580-af0eb2d31ef8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.385129 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c569fb8-edf7-4012-bd27-3eb30c568489-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3c569fb8-edf7-4012-bd27-3eb30c568489" (UID: "3c569fb8-edf7-4012-bd27-3eb30c568489"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.387515 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c569fb8-edf7-4012-bd27-3eb30c568489-kube-api-access-j7jn4" (OuterVolumeSpecName: "kube-api-access-j7jn4") pod "3c569fb8-edf7-4012-bd27-3eb30c568489" (UID: "3c569fb8-edf7-4012-bd27-3eb30c568489"). InnerVolumeSpecName "kube-api-access-j7jn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.388070 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d5470b1-6bce-4dac-b1a7-f2ed50bcb713-kube-api-access-j7qd2" (OuterVolumeSpecName: "kube-api-access-j7qd2") pod "0d5470b1-6bce-4dac-b1a7-f2ed50bcb713" (UID: "0d5470b1-6bce-4dac-b1a7-f2ed50bcb713"). InnerVolumeSpecName "kube-api-access-j7qd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.388141 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7067c05-68d2-413d-9580-af0eb2d31ef8-kube-api-access-bhzk7" (OuterVolumeSpecName: "kube-api-access-bhzk7") pod "c7067c05-68d2-413d-9580-af0eb2d31ef8" (UID: "c7067c05-68d2-413d-9580-af0eb2d31ef8"). InnerVolumeSpecName "kube-api-access-bhzk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.485838 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0936beb9-f71b-4313-b6a0-97b026f697a5-operator-scripts\") pod \"0936beb9-f71b-4313-b6a0-97b026f697a5\" (UID: \"0936beb9-f71b-4313-b6a0-97b026f697a5\") " Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.485912 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5ch4\" (UniqueName: \"kubernetes.io/projected/0227f04c-42f9-4a76-bd79-c1702200ca30-kube-api-access-z5ch4\") pod \"0227f04c-42f9-4a76-bd79-c1702200ca30\" (UID: \"0227f04c-42f9-4a76-bd79-c1702200ca30\") " Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.485980 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0227f04c-42f9-4a76-bd79-c1702200ca30-operator-scripts\") pod \"0227f04c-42f9-4a76-bd79-c1702200ca30\" (UID: \"0227f04c-42f9-4a76-bd79-c1702200ca30\") " Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.486021 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fj2vq\" (UniqueName: \"kubernetes.io/projected/0936beb9-f71b-4313-b6a0-97b026f697a5-kube-api-access-fj2vq\") pod \"0936beb9-f71b-4313-b6a0-97b026f697a5\" (UID: \"0936beb9-f71b-4313-b6a0-97b026f697a5\") " Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.486597 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0936beb9-f71b-4313-b6a0-97b026f697a5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0936beb9-f71b-4313-b6a0-97b026f697a5" (UID: "0936beb9-f71b-4313-b6a0-97b026f697a5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.486628 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0227f04c-42f9-4a76-bd79-c1702200ca30-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0227f04c-42f9-4a76-bd79-c1702200ca30" (UID: "0227f04c-42f9-4a76-bd79-c1702200ca30"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.487203 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0227f04c-42f9-4a76-bd79-c1702200ca30-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.487223 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c569fb8-edf7-4012-bd27-3eb30c568489-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.487233 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7qd2\" (UniqueName: \"kubernetes.io/projected/0d5470b1-6bce-4dac-b1a7-f2ed50bcb713-kube-api-access-j7qd2\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.487244 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7067c05-68d2-413d-9580-af0eb2d31ef8-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.487309 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0936beb9-f71b-4313-b6a0-97b026f697a5-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.487320 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhzk7\" (UniqueName: \"kubernetes.io/projected/c7067c05-68d2-413d-9580-af0eb2d31ef8-kube-api-access-bhzk7\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.487328 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7jn4\" (UniqueName: \"kubernetes.io/projected/3c569fb8-edf7-4012-bd27-3eb30c568489-kube-api-access-j7jn4\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.489190 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0227f04c-42f9-4a76-bd79-c1702200ca30-kube-api-access-z5ch4" (OuterVolumeSpecName: "kube-api-access-z5ch4") pod "0227f04c-42f9-4a76-bd79-c1702200ca30" (UID: "0227f04c-42f9-4a76-bd79-c1702200ca30"). InnerVolumeSpecName "kube-api-access-z5ch4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.489778 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0936beb9-f71b-4313-b6a0-97b026f697a5-kube-api-access-fj2vq" (OuterVolumeSpecName: "kube-api-access-fj2vq") pod "0936beb9-f71b-4313-b6a0-97b026f697a5" (UID: "0936beb9-f71b-4313-b6a0-97b026f697a5"). InnerVolumeSpecName "kube-api-access-fj2vq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.589379 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5ch4\" (UniqueName: \"kubernetes.io/projected/0227f04c-42f9-4a76-bd79-c1702200ca30-kube-api-access-z5ch4\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.589412 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fj2vq\" (UniqueName: \"kubernetes.io/projected/0936beb9-f71b-4313-b6a0-97b026f697a5-kube-api-access-fj2vq\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.709670 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-7fs72" event={"ID":"0936beb9-f71b-4313-b6a0-97b026f697a5","Type":"ContainerDied","Data":"2f929285f063c9e68a989915acb4cd7824c5e44d04924c3682fe8299dcd1f701"} Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.709708 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f929285f063c9e68a989915acb4cd7824c5e44d04924c3682fe8299dcd1f701" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.709707 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-7fs72" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.711189 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-b2hh7" event={"ID":"3c569fb8-edf7-4012-bd27-3eb30c568489","Type":"ContainerDied","Data":"52905bc118d72d16fff3ae82209e42e1c5c480b0af9606bf6b9424cadb980eeb"} Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.711206 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-b2hh7" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.711224 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52905bc118d72d16fff3ae82209e42e1c5c480b0af9606bf6b9424cadb980eeb" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.712559 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-nghsz" event={"ID":"c7067c05-68d2-413d-9580-af0eb2d31ef8","Type":"ContainerDied","Data":"21bb64f3b4ae4bbae94f33fad224d256b7f4b1364d493584028ef9d91784db22"} Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.712672 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21bb64f3b4ae4bbae94f33fad224d256b7f4b1364d493584028ef9d91784db22" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.712577 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-nghsz" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.713950 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-552e-account-create-update-t4h4f" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.713955 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-552e-account-create-update-t4h4f" event={"ID":"0227f04c-42f9-4a76-bd79-c1702200ca30","Type":"ContainerDied","Data":"c47951d0a611160aeb80af4ed58e6a20060a5db2360b2e2227b3f47a16aa51cd"} Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.714493 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c47951d0a611160aeb80af4ed58e6a20060a5db2360b2e2227b3f47a16aa51cd" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.715858 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d153-account-create-update-728q9" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.715980 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d153-account-create-update-728q9" event={"ID":"90f4f716-4da8-4ab0-8757-1c15480c73b5","Type":"ContainerDied","Data":"0e08943a2c58e15cdc71e51cf5c443f34068fa535cf3e741dcaa9978f5150e2d"} Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.716016 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e08943a2c58e15cdc71e51cf5c443f34068fa535cf3e741dcaa9978f5150e2d" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.717746 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-4f6e-account-create-update-wx2s5" event={"ID":"0d5470b1-6bce-4dac-b1a7-f2ed50bcb713","Type":"ContainerDied","Data":"609243d58967760c92dc69a36f089e3f7e0f43830e4eeee43ccfe1b775dc767e"} Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.717876 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4f6e-account-create-update-wx2s5" Mar 21 08:46:24 crc kubenswrapper[4696]: I0321 08:46:24.717864 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="609243d58967760c92dc69a36f089e3f7e0f43830e4eeee43ccfe1b775dc767e" Mar 21 08:46:25 crc kubenswrapper[4696]: I0321 08:46:25.728557 4696 generic.go:334] "Generic (PLEG): container finished" podID="b057ad21-8030-49e9-b400-c36d433b9f8c" containerID="7fd27604a7d301ac11fa0074a203113325187e8f947c4308c11343eb67ceaed5" exitCode=0 Mar 21 08:46:25 crc kubenswrapper[4696]: I0321 08:46:25.728611 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b057ad21-8030-49e9-b400-c36d433b9f8c","Type":"ContainerDied","Data":"7fd27604a7d301ac11fa0074a203113325187e8f947c4308c11343eb67ceaed5"} Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.224682 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-vgptg"] Mar 21 08:46:26 crc kubenswrapper[4696]: E0321 08:46:26.225291 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efd827ca-ae90-4cca-a77d-72fce76c47aa" containerName="swift-ring-rebalance" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.225309 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="efd827ca-ae90-4cca-a77d-72fce76c47aa" containerName="swift-ring-rebalance" Mar 21 08:46:26 crc kubenswrapper[4696]: E0321 08:46:26.225330 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90f4f716-4da8-4ab0-8757-1c15480c73b5" containerName="mariadb-account-create-update" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.225337 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="90f4f716-4da8-4ab0-8757-1c15480c73b5" containerName="mariadb-account-create-update" Mar 21 08:46:26 crc kubenswrapper[4696]: E0321 08:46:26.225345 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d5470b1-6bce-4dac-b1a7-f2ed50bcb713" containerName="mariadb-account-create-update" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.225352 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d5470b1-6bce-4dac-b1a7-f2ed50bcb713" containerName="mariadb-account-create-update" Mar 21 08:46:26 crc kubenswrapper[4696]: E0321 08:46:26.225364 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c569fb8-edf7-4012-bd27-3eb30c568489" containerName="mariadb-database-create" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.225372 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c569fb8-edf7-4012-bd27-3eb30c568489" containerName="mariadb-database-create" Mar 21 08:46:26 crc kubenswrapper[4696]: E0321 08:46:26.225383 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7067c05-68d2-413d-9580-af0eb2d31ef8" containerName="mariadb-database-create" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.225391 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7067c05-68d2-413d-9580-af0eb2d31ef8" containerName="mariadb-database-create" Mar 21 08:46:26 crc kubenswrapper[4696]: E0321 08:46:26.225407 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0936beb9-f71b-4313-b6a0-97b026f697a5" containerName="mariadb-database-create" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.225415 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0936beb9-f71b-4313-b6a0-97b026f697a5" containerName="mariadb-database-create" Mar 21 08:46:26 crc kubenswrapper[4696]: E0321 08:46:26.225435 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0227f04c-42f9-4a76-bd79-c1702200ca30" containerName="mariadb-account-create-update" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.225443 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0227f04c-42f9-4a76-bd79-c1702200ca30" containerName="mariadb-account-create-update" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.225649 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d5470b1-6bce-4dac-b1a7-f2ed50bcb713" containerName="mariadb-account-create-update" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.225663 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c569fb8-edf7-4012-bd27-3eb30c568489" containerName="mariadb-database-create" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.225671 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7067c05-68d2-413d-9580-af0eb2d31ef8" containerName="mariadb-database-create" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.225687 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="efd827ca-ae90-4cca-a77d-72fce76c47aa" containerName="swift-ring-rebalance" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.225697 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="0936beb9-f71b-4313-b6a0-97b026f697a5" containerName="mariadb-database-create" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.225705 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="0227f04c-42f9-4a76-bd79-c1702200ca30" containerName="mariadb-account-create-update" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.225718 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="90f4f716-4da8-4ab0-8757-1c15480c73b5" containerName="mariadb-account-create-update" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.226573 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-vgptg" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.229030 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.234688 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-vgptg"] Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.316932 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1-operator-scripts\") pod \"root-account-create-update-vgptg\" (UID: \"5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1\") " pod="openstack/root-account-create-update-vgptg" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.317013 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vxqk\" (UniqueName: \"kubernetes.io/projected/5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1-kube-api-access-9vxqk\") pod \"root-account-create-update-vgptg\" (UID: \"5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1\") " pod="openstack/root-account-create-update-vgptg" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.418271 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vxqk\" (UniqueName: \"kubernetes.io/projected/5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1-kube-api-access-9vxqk\") pod \"root-account-create-update-vgptg\" (UID: \"5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1\") " pod="openstack/root-account-create-update-vgptg" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.418432 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1-operator-scripts\") pod \"root-account-create-update-vgptg\" (UID: \"5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1\") " pod="openstack/root-account-create-update-vgptg" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.419337 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1-operator-scripts\") pod \"root-account-create-update-vgptg\" (UID: \"5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1\") " pod="openstack/root-account-create-update-vgptg" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.437936 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vxqk\" (UniqueName: \"kubernetes.io/projected/5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1-kube-api-access-9vxqk\") pod \"root-account-create-update-vgptg\" (UID: \"5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1\") " pod="openstack/root-account-create-update-vgptg" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.543241 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-vgptg" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.741518 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b057ad21-8030-49e9-b400-c36d433b9f8c","Type":"ContainerStarted","Data":"b9bb5d044956676c0e0405393dd7743a5061cd8b121716b8f1aa7cd1c1987a17"} Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.742266 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.743849 4696 generic.go:334] "Generic (PLEG): container finished" podID="0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" containerID="e8e800b9e6d78c36e85ad2b994e32f91f76eb2ffdc75250bf0f1b501cb1813a8" exitCode=0 Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.743878 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a","Type":"ContainerDied","Data":"e8e800b9e6d78c36e85ad2b994e32f91f76eb2ffdc75250bf0f1b501cb1813a8"} Mar 21 08:46:26 crc kubenswrapper[4696]: I0321 08:46:26.811503 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=56.171271874 podStartE2EDuration="1m12.811475786s" podCreationTimestamp="2026-03-21 08:45:14 +0000 UTC" firstStartedPulling="2026-03-21 08:45:33.914076242 +0000 UTC m=+1068.034956955" lastFinishedPulling="2026-03-21 08:45:50.554280154 +0000 UTC m=+1084.675160867" observedRunningTime="2026-03-21 08:46:26.803106322 +0000 UTC m=+1120.923987035" watchObservedRunningTime="2026-03-21 08:46:26.811475786 +0000 UTC m=+1120.932356499" Mar 21 08:46:27 crc kubenswrapper[4696]: W0321 08:46:27.102965 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ccfa6bc_c97b_4b79_b8cc_466cb3e676b1.slice/crio-1e1693dfb3f0cdea753cd28a5502921a7896a2f3488661c85f8870a8a5ccc4b7 WatchSource:0}: Error finding container 1e1693dfb3f0cdea753cd28a5502921a7896a2f3488661c85f8870a8a5ccc4b7: Status 404 returned error can't find the container with id 1e1693dfb3f0cdea753cd28a5502921a7896a2f3488661c85f8870a8a5ccc4b7 Mar 21 08:46:27 crc kubenswrapper[4696]: I0321 08:46:27.105952 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-vgptg"] Mar 21 08:46:27 crc kubenswrapper[4696]: I0321 08:46:27.543910 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 21 08:46:27 crc kubenswrapper[4696]: I0321 08:46:27.544163 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" containerName="prometheus" containerID="cri-o://6c31373671bc55e8b663291df3d6014ea1784bed879432d8e143689cd7b1d5b8" gracePeriod=600 Mar 21 08:46:27 crc kubenswrapper[4696]: I0321 08:46:27.544220 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" containerName="config-reloader" containerID="cri-o://45bca271f037081231d7d8df544b677a84f380931a1ae8f272313aec3e1f6097" gracePeriod=600 Mar 21 08:46:27 crc kubenswrapper[4696]: I0321 08:46:27.544229 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" containerName="thanos-sidecar" containerID="cri-o://10d17fe0abf4d917f53465a679a5fbdd91f1967736cd68e4446a6853891177b2" gracePeriod=600 Mar 21 08:46:27 crc kubenswrapper[4696]: I0321 08:46:27.754376 4696 generic.go:334] "Generic (PLEG): container finished" podID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" containerID="10d17fe0abf4d917f53465a679a5fbdd91f1967736cd68e4446a6853891177b2" exitCode=0 Mar 21 08:46:27 crc kubenswrapper[4696]: I0321 08:46:27.754405 4696 generic.go:334] "Generic (PLEG): container finished" podID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" containerID="6c31373671bc55e8b663291df3d6014ea1784bed879432d8e143689cd7b1d5b8" exitCode=0 Mar 21 08:46:27 crc kubenswrapper[4696]: I0321 08:46:27.754450 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5b9177a1-3d3a-4a87-a1fa-1dbe96439288","Type":"ContainerDied","Data":"10d17fe0abf4d917f53465a679a5fbdd91f1967736cd68e4446a6853891177b2"} Mar 21 08:46:27 crc kubenswrapper[4696]: I0321 08:46:27.754499 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5b9177a1-3d3a-4a87-a1fa-1dbe96439288","Type":"ContainerDied","Data":"6c31373671bc55e8b663291df3d6014ea1784bed879432d8e143689cd7b1d5b8"} Mar 21 08:46:27 crc kubenswrapper[4696]: I0321 08:46:27.757713 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a","Type":"ContainerStarted","Data":"8be3270ef4edb543fbfd2456b063317996cad6cb123ab298417ce81d8c7a08c6"} Mar 21 08:46:27 crc kubenswrapper[4696]: I0321 08:46:27.758188 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 21 08:46:27 crc kubenswrapper[4696]: I0321 08:46:27.761879 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-vgptg" event={"ID":"5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1","Type":"ContainerStarted","Data":"ebe25e48963bc833c63e1295302833398482ec0ae4b9f85c9c5e746501f31a64"} Mar 21 08:46:27 crc kubenswrapper[4696]: I0321 08:46:27.761930 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-vgptg" event={"ID":"5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1","Type":"ContainerStarted","Data":"1e1693dfb3f0cdea753cd28a5502921a7896a2f3488661c85f8870a8a5ccc4b7"} Mar 21 08:46:27 crc kubenswrapper[4696]: I0321 08:46:27.790947 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=56.986979702 podStartE2EDuration="1m12.790928287s" podCreationTimestamp="2026-03-21 08:45:15 +0000 UTC" firstStartedPulling="2026-03-21 08:45:35.128042622 +0000 UTC m=+1069.248923335" lastFinishedPulling="2026-03-21 08:45:50.931991207 +0000 UTC m=+1085.052871920" observedRunningTime="2026-03-21 08:46:27.782920444 +0000 UTC m=+1121.903801167" watchObservedRunningTime="2026-03-21 08:46:27.790928287 +0000 UTC m=+1121.911809000" Mar 21 08:46:27 crc kubenswrapper[4696]: I0321 08:46:27.805456 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-vgptg" podStartSLOduration=1.805434243 podStartE2EDuration="1.805434243s" podCreationTimestamp="2026-03-21 08:46:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:46:27.803386465 +0000 UTC m=+1121.924267178" watchObservedRunningTime="2026-03-21 08:46:27.805434243 +0000 UTC m=+1121.926314956" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.189120 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.116:9090/-/ready\": dial tcp 10.217.0.116:9090: connect: connection refused" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.530020 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.664549 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lqpr\" (UniqueName: \"kubernetes.io/projected/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-kube-api-access-9lqpr\") pod \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.664696 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-thanos-prometheus-http-client-file\") pod \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.664764 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-config-out\") pod \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.664800 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-prometheus-metric-storage-rulefiles-2\") pod \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.664873 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-prometheus-metric-storage-rulefiles-0\") pod \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.664956 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-config\") pod \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.665365 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "5b9177a1-3d3a-4a87-a1fa-1dbe96439288" (UID: "5b9177a1-3d3a-4a87-a1fa-1dbe96439288"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.664982 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-prometheus-metric-storage-rulefiles-1\") pod \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.666185 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-prometheus-metric-storage-rulefiles-2" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-2") pod "5b9177a1-3d3a-4a87-a1fa-1dbe96439288" (UID: "5b9177a1-3d3a-4a87-a1fa-1dbe96439288"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-2". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.666677 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-prometheus-metric-storage-rulefiles-1" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-1") pod "5b9177a1-3d3a-4a87-a1fa-1dbe96439288" (UID: "5b9177a1-3d3a-4a87-a1fa-1dbe96439288"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.666997 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-tls-assets\") pod \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.667026 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-web-config\") pod \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.667144 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af131abf-8240-4971-8fb5-3805d310c306\") pod \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\" (UID: \"5b9177a1-3d3a-4a87-a1fa-1dbe96439288\") " Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.667926 4696 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-prometheus-metric-storage-rulefiles-2\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.667944 4696 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.667955 4696 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-prometheus-metric-storage-rulefiles-1\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.671036 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "5b9177a1-3d3a-4a87-a1fa-1dbe96439288" (UID: "5b9177a1-3d3a-4a87-a1fa-1dbe96439288"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.677972 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "5b9177a1-3d3a-4a87-a1fa-1dbe96439288" (UID: "5b9177a1-3d3a-4a87-a1fa-1dbe96439288"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.678067 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-config" (OuterVolumeSpecName: "config") pod "5b9177a1-3d3a-4a87-a1fa-1dbe96439288" (UID: "5b9177a1-3d3a-4a87-a1fa-1dbe96439288"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.678219 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-kube-api-access-9lqpr" (OuterVolumeSpecName: "kube-api-access-9lqpr") pod "5b9177a1-3d3a-4a87-a1fa-1dbe96439288" (UID: "5b9177a1-3d3a-4a87-a1fa-1dbe96439288"). InnerVolumeSpecName "kube-api-access-9lqpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.691049 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-web-config" (OuterVolumeSpecName: "web-config") pod "5b9177a1-3d3a-4a87-a1fa-1dbe96439288" (UID: "5b9177a1-3d3a-4a87-a1fa-1dbe96439288"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.706236 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-config-out" (OuterVolumeSpecName: "config-out") pod "5b9177a1-3d3a-4a87-a1fa-1dbe96439288" (UID: "5b9177a1-3d3a-4a87-a1fa-1dbe96439288"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.712676 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af131abf-8240-4971-8fb5-3805d310c306" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "5b9177a1-3d3a-4a87-a1fa-1dbe96439288" (UID: "5b9177a1-3d3a-4a87-a1fa-1dbe96439288"). InnerVolumeSpecName "pvc-af131abf-8240-4971-8fb5-3805d310c306". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.771676 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.771723 4696 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-tls-assets\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.771742 4696 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-web-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.771790 4696 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-af131abf-8240-4971-8fb5-3805d310c306\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af131abf-8240-4971-8fb5-3805d310c306\") on node \"crc\" " Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.771867 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lqpr\" (UniqueName: \"kubernetes.io/projected/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-kube-api-access-9lqpr\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.771891 4696 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.771910 4696 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5b9177a1-3d3a-4a87-a1fa-1dbe96439288-config-out\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.780703 4696 generic.go:334] "Generic (PLEG): container finished" podID="5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1" containerID="ebe25e48963bc833c63e1295302833398482ec0ae4b9f85c9c5e746501f31a64" exitCode=0 Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.780752 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-vgptg" event={"ID":"5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1","Type":"ContainerDied","Data":"ebe25e48963bc833c63e1295302833398482ec0ae4b9f85c9c5e746501f31a64"} Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.783840 4696 generic.go:334] "Generic (PLEG): container finished" podID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" containerID="45bca271f037081231d7d8df544b677a84f380931a1ae8f272313aec3e1f6097" exitCode=0 Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.783881 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.783940 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5b9177a1-3d3a-4a87-a1fa-1dbe96439288","Type":"ContainerDied","Data":"45bca271f037081231d7d8df544b677a84f380931a1ae8f272313aec3e1f6097"} Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.783978 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5b9177a1-3d3a-4a87-a1fa-1dbe96439288","Type":"ContainerDied","Data":"fa2c7c7d410cc660805060a6002d5068a2194c994a8895b165f06b0b23a313d7"} Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.784000 4696 scope.go:117] "RemoveContainer" containerID="10d17fe0abf4d917f53465a679a5fbdd91f1967736cd68e4446a6853891177b2" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.795130 4696 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.795263 4696 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-af131abf-8240-4971-8fb5-3805d310c306" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af131abf-8240-4971-8fb5-3805d310c306") on node "crc" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.808677 4696 scope.go:117] "RemoveContainer" containerID="45bca271f037081231d7d8df544b677a84f380931a1ae8f272313aec3e1f6097" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.840860 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.841465 4696 scope.go:117] "RemoveContainer" containerID="6c31373671bc55e8b663291df3d6014ea1784bed879432d8e143689cd7b1d5b8" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.853685 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.866757 4696 scope.go:117] "RemoveContainer" containerID="853f5cfbde31fd213428791fc1e517f055f22c0aca0b293c96501d865a534f07" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.868378 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 21 08:46:28 crc kubenswrapper[4696]: E0321 08:46:28.868793 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" containerName="config-reloader" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.868835 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" containerName="config-reloader" Mar 21 08:46:28 crc kubenswrapper[4696]: E0321 08:46:28.868859 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" containerName="thanos-sidecar" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.868867 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" containerName="thanos-sidecar" Mar 21 08:46:28 crc kubenswrapper[4696]: E0321 08:46:28.868889 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" containerName="prometheus" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.868896 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" containerName="prometheus" Mar 21 08:46:28 crc kubenswrapper[4696]: E0321 08:46:28.868909 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" containerName="init-config-reloader" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.868917 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" containerName="init-config-reloader" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.869120 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" containerName="config-reloader" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.869150 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" containerName="prometheus" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.869162 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" containerName="thanos-sidecar" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.871199 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.874873 4696 reconciler_common.go:293] "Volume detached for volume \"pvc-af131abf-8240-4971-8fb5-3805d310c306\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af131abf-8240-4971-8fb5-3805d310c306\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.876012 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.876290 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-1" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.877873 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.877887 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-v6ljh" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.877887 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.877968 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.878014 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-2" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.878150 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.882841 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.889074 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.964998 4696 scope.go:117] "RemoveContainer" containerID="10d17fe0abf4d917f53465a679a5fbdd91f1967736cd68e4446a6853891177b2" Mar 21 08:46:28 crc kubenswrapper[4696]: E0321 08:46:28.969301 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10d17fe0abf4d917f53465a679a5fbdd91f1967736cd68e4446a6853891177b2\": container with ID starting with 10d17fe0abf4d917f53465a679a5fbdd91f1967736cd68e4446a6853891177b2 not found: ID does not exist" containerID="10d17fe0abf4d917f53465a679a5fbdd91f1967736cd68e4446a6853891177b2" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.969347 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10d17fe0abf4d917f53465a679a5fbdd91f1967736cd68e4446a6853891177b2"} err="failed to get container status \"10d17fe0abf4d917f53465a679a5fbdd91f1967736cd68e4446a6853891177b2\": rpc error: code = NotFound desc = could not find container \"10d17fe0abf4d917f53465a679a5fbdd91f1967736cd68e4446a6853891177b2\": container with ID starting with 10d17fe0abf4d917f53465a679a5fbdd91f1967736cd68e4446a6853891177b2 not found: ID does not exist" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.969378 4696 scope.go:117] "RemoveContainer" containerID="45bca271f037081231d7d8df544b677a84f380931a1ae8f272313aec3e1f6097" Mar 21 08:46:28 crc kubenswrapper[4696]: E0321 08:46:28.970101 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45bca271f037081231d7d8df544b677a84f380931a1ae8f272313aec3e1f6097\": container with ID starting with 45bca271f037081231d7d8df544b677a84f380931a1ae8f272313aec3e1f6097 not found: ID does not exist" containerID="45bca271f037081231d7d8df544b677a84f380931a1ae8f272313aec3e1f6097" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.970139 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45bca271f037081231d7d8df544b677a84f380931a1ae8f272313aec3e1f6097"} err="failed to get container status \"45bca271f037081231d7d8df544b677a84f380931a1ae8f272313aec3e1f6097\": rpc error: code = NotFound desc = could not find container \"45bca271f037081231d7d8df544b677a84f380931a1ae8f272313aec3e1f6097\": container with ID starting with 45bca271f037081231d7d8df544b677a84f380931a1ae8f272313aec3e1f6097 not found: ID does not exist" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.970157 4696 scope.go:117] "RemoveContainer" containerID="6c31373671bc55e8b663291df3d6014ea1784bed879432d8e143689cd7b1d5b8" Mar 21 08:46:28 crc kubenswrapper[4696]: E0321 08:46:28.971062 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c31373671bc55e8b663291df3d6014ea1784bed879432d8e143689cd7b1d5b8\": container with ID starting with 6c31373671bc55e8b663291df3d6014ea1784bed879432d8e143689cd7b1d5b8 not found: ID does not exist" containerID="6c31373671bc55e8b663291df3d6014ea1784bed879432d8e143689cd7b1d5b8" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.971168 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c31373671bc55e8b663291df3d6014ea1784bed879432d8e143689cd7b1d5b8"} err="failed to get container status \"6c31373671bc55e8b663291df3d6014ea1784bed879432d8e143689cd7b1d5b8\": rpc error: code = NotFound desc = could not find container \"6c31373671bc55e8b663291df3d6014ea1784bed879432d8e143689cd7b1d5b8\": container with ID starting with 6c31373671bc55e8b663291df3d6014ea1784bed879432d8e143689cd7b1d5b8 not found: ID does not exist" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.971257 4696 scope.go:117] "RemoveContainer" containerID="853f5cfbde31fd213428791fc1e517f055f22c0aca0b293c96501d865a534f07" Mar 21 08:46:28 crc kubenswrapper[4696]: E0321 08:46:28.972737 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"853f5cfbde31fd213428791fc1e517f055f22c0aca0b293c96501d865a534f07\": container with ID starting with 853f5cfbde31fd213428791fc1e517f055f22c0aca0b293c96501d865a534f07 not found: ID does not exist" containerID="853f5cfbde31fd213428791fc1e517f055f22c0aca0b293c96501d865a534f07" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.972767 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"853f5cfbde31fd213428791fc1e517f055f22c0aca0b293c96501d865a534f07"} err="failed to get container status \"853f5cfbde31fd213428791fc1e517f055f22c0aca0b293c96501d865a534f07\": rpc error: code = NotFound desc = could not find container \"853f5cfbde31fd213428791fc1e517f055f22c0aca0b293c96501d865a534f07\": container with ID starting with 853f5cfbde31fd213428791fc1e517f055f22c0aca0b293c96501d865a534f07 not found: ID does not exist" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.976345 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c2c63b45-de34-43a6-bb8b-aed4b00922c8-config\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.976399 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/c2c63b45-de34-43a6-bb8b-aed4b00922c8-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.976439 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-af131abf-8240-4971-8fb5-3805d310c306\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af131abf-8240-4971-8fb5-3805d310c306\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.976490 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/c2c63b45-de34-43a6-bb8b-aed4b00922c8-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.976526 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84rkj\" (UniqueName: \"kubernetes.io/projected/c2c63b45-de34-43a6-bb8b-aed4b00922c8-kube-api-access-84rkj\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.976569 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2c63b45-de34-43a6-bb8b-aed4b00922c8-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.976628 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/c2c63b45-de34-43a6-bb8b-aed4b00922c8-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.976648 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/c2c63b45-de34-43a6-bb8b-aed4b00922c8-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.976665 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/c2c63b45-de34-43a6-bb8b-aed4b00922c8-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.976690 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/c2c63b45-de34-43a6-bb8b-aed4b00922c8-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.976709 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/c2c63b45-de34-43a6-bb8b-aed4b00922c8-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.976729 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/c2c63b45-de34-43a6-bb8b-aed4b00922c8-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:28 crc kubenswrapper[4696]: I0321 08:46:28.977197 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/c2c63b45-de34-43a6-bb8b-aed4b00922c8-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.078666 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/c2c63b45-de34-43a6-bb8b-aed4b00922c8-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.078726 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84rkj\" (UniqueName: \"kubernetes.io/projected/c2c63b45-de34-43a6-bb8b-aed4b00922c8-kube-api-access-84rkj\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.078774 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2c63b45-de34-43a6-bb8b-aed4b00922c8-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.078850 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/c2c63b45-de34-43a6-bb8b-aed4b00922c8-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.078870 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/c2c63b45-de34-43a6-bb8b-aed4b00922c8-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.078887 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/c2c63b45-de34-43a6-bb8b-aed4b00922c8-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.078914 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/c2c63b45-de34-43a6-bb8b-aed4b00922c8-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.078934 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/c2c63b45-de34-43a6-bb8b-aed4b00922c8-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.078953 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/c2c63b45-de34-43a6-bb8b-aed4b00922c8-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.078970 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/c2c63b45-de34-43a6-bb8b-aed4b00922c8-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.078996 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c2c63b45-de34-43a6-bb8b-aed4b00922c8-config\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.079021 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/c2c63b45-de34-43a6-bb8b-aed4b00922c8-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.079040 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-af131abf-8240-4971-8fb5-3805d310c306\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af131abf-8240-4971-8fb5-3805d310c306\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.080201 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/c2c63b45-de34-43a6-bb8b-aed4b00922c8-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.080529 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/c2c63b45-de34-43a6-bb8b-aed4b00922c8-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.080682 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/c2c63b45-de34-43a6-bb8b-aed4b00922c8-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.083006 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/c2c63b45-de34-43a6-bb8b-aed4b00922c8-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.084126 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/c2c63b45-de34-43a6-bb8b-aed4b00922c8-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.084582 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2c63b45-de34-43a6-bb8b-aed4b00922c8-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.084825 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/c2c63b45-de34-43a6-bb8b-aed4b00922c8-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.085289 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c2c63b45-de34-43a6-bb8b-aed4b00922c8-config\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.086291 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/c2c63b45-de34-43a6-bb8b-aed4b00922c8-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.087059 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/c2c63b45-de34-43a6-bb8b-aed4b00922c8-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.088578 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/c2c63b45-de34-43a6-bb8b-aed4b00922c8-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.089164 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.089202 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-af131abf-8240-4971-8fb5-3805d310c306\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af131abf-8240-4971-8fb5-3805d310c306\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d67ff30a3d5f27596f87a4e17cfdebdac6e352f692a21f0c4f6bda3eb4f6d98f/globalmount\"" pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.099793 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84rkj\" (UniqueName: \"kubernetes.io/projected/c2c63b45-de34-43a6-bb8b-aed4b00922c8-kube-api-access-84rkj\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.128611 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-af131abf-8240-4971-8fb5-3805d310c306\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af131abf-8240-4971-8fb5-3805d310c306\") pod \"prometheus-metric-storage-0\" (UID: \"c2c63b45-de34-43a6-bb8b-aed4b00922c8\") " pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.283680 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.565341 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-jqpz7"] Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.567365 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jqpz7" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.575068 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-f5cmp" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.581317 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-jqpz7"] Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.582430 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.688444 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7359c71c-56ba-4c28-a126-bc95dea18dfd-combined-ca-bundle\") pod \"glance-db-sync-jqpz7\" (UID: \"7359c71c-56ba-4c28-a126-bc95dea18dfd\") " pod="openstack/glance-db-sync-jqpz7" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.688603 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmb9h\" (UniqueName: \"kubernetes.io/projected/7359c71c-56ba-4c28-a126-bc95dea18dfd-kube-api-access-kmb9h\") pod \"glance-db-sync-jqpz7\" (UID: \"7359c71c-56ba-4c28-a126-bc95dea18dfd\") " pod="openstack/glance-db-sync-jqpz7" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.688626 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7359c71c-56ba-4c28-a126-bc95dea18dfd-config-data\") pod \"glance-db-sync-jqpz7\" (UID: \"7359c71c-56ba-4c28-a126-bc95dea18dfd\") " pod="openstack/glance-db-sync-jqpz7" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.688660 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7359c71c-56ba-4c28-a126-bc95dea18dfd-db-sync-config-data\") pod \"glance-db-sync-jqpz7\" (UID: \"7359c71c-56ba-4c28-a126-bc95dea18dfd\") " pod="openstack/glance-db-sync-jqpz7" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.773348 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.790503 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmb9h\" (UniqueName: \"kubernetes.io/projected/7359c71c-56ba-4c28-a126-bc95dea18dfd-kube-api-access-kmb9h\") pod \"glance-db-sync-jqpz7\" (UID: \"7359c71c-56ba-4c28-a126-bc95dea18dfd\") " pod="openstack/glance-db-sync-jqpz7" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.790542 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7359c71c-56ba-4c28-a126-bc95dea18dfd-config-data\") pod \"glance-db-sync-jqpz7\" (UID: \"7359c71c-56ba-4c28-a126-bc95dea18dfd\") " pod="openstack/glance-db-sync-jqpz7" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.790580 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7359c71c-56ba-4c28-a126-bc95dea18dfd-db-sync-config-data\") pod \"glance-db-sync-jqpz7\" (UID: \"7359c71c-56ba-4c28-a126-bc95dea18dfd\") " pod="openstack/glance-db-sync-jqpz7" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.790612 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7359c71c-56ba-4c28-a126-bc95dea18dfd-combined-ca-bundle\") pod \"glance-db-sync-jqpz7\" (UID: \"7359c71c-56ba-4c28-a126-bc95dea18dfd\") " pod="openstack/glance-db-sync-jqpz7" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.796431 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7359c71c-56ba-4c28-a126-bc95dea18dfd-combined-ca-bundle\") pod \"glance-db-sync-jqpz7\" (UID: \"7359c71c-56ba-4c28-a126-bc95dea18dfd\") " pod="openstack/glance-db-sync-jqpz7" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.800145 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7359c71c-56ba-4c28-a126-bc95dea18dfd-db-sync-config-data\") pod \"glance-db-sync-jqpz7\" (UID: \"7359c71c-56ba-4c28-a126-bc95dea18dfd\") " pod="openstack/glance-db-sync-jqpz7" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.804594 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7359c71c-56ba-4c28-a126-bc95dea18dfd-config-data\") pod \"glance-db-sync-jqpz7\" (UID: \"7359c71c-56ba-4c28-a126-bc95dea18dfd\") " pod="openstack/glance-db-sync-jqpz7" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.817771 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmb9h\" (UniqueName: \"kubernetes.io/projected/7359c71c-56ba-4c28-a126-bc95dea18dfd-kube-api-access-kmb9h\") pod \"glance-db-sync-jqpz7\" (UID: \"7359c71c-56ba-4c28-a126-bc95dea18dfd\") " pod="openstack/glance-db-sync-jqpz7" Mar 21 08:46:29 crc kubenswrapper[4696]: I0321 08:46:29.892113 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jqpz7" Mar 21 08:46:30 crc kubenswrapper[4696]: I0321 08:46:30.398428 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-vgptg" Mar 21 08:46:30 crc kubenswrapper[4696]: I0321 08:46:30.504087 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vxqk\" (UniqueName: \"kubernetes.io/projected/5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1-kube-api-access-9vxqk\") pod \"5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1\" (UID: \"5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1\") " Mar 21 08:46:30 crc kubenswrapper[4696]: I0321 08:46:30.504191 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1-operator-scripts\") pod \"5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1\" (UID: \"5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1\") " Mar 21 08:46:30 crc kubenswrapper[4696]: I0321 08:46:30.504921 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1" (UID: "5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:30 crc kubenswrapper[4696]: I0321 08:46:30.505323 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:30 crc kubenswrapper[4696]: I0321 08:46:30.510350 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1-kube-api-access-9vxqk" (OuterVolumeSpecName: "kube-api-access-9vxqk") pod "5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1" (UID: "5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1"). InnerVolumeSpecName "kube-api-access-9vxqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:30 crc kubenswrapper[4696]: I0321 08:46:30.545303 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b9177a1-3d3a-4a87-a1fa-1dbe96439288" path="/var/lib/kubelet/pods/5b9177a1-3d3a-4a87-a1fa-1dbe96439288/volumes" Mar 21 08:46:30 crc kubenswrapper[4696]: I0321 08:46:30.607007 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vxqk\" (UniqueName: \"kubernetes.io/projected/5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1-kube-api-access-9vxqk\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:30 crc kubenswrapper[4696]: I0321 08:46:30.625237 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-jqpz7"] Mar 21 08:46:30 crc kubenswrapper[4696]: I0321 08:46:30.656752 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="d6518eba-34ec-4a38-b279-cd8bb11e7e24" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 21 08:46:30 crc kubenswrapper[4696]: I0321 08:46:30.815810 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jqpz7" event={"ID":"7359c71c-56ba-4c28-a126-bc95dea18dfd","Type":"ContainerStarted","Data":"33f6e2facaef341cf101f0a2eb4651da0a268cdb95c8f7d9e9238691912feb19"} Mar 21 08:46:30 crc kubenswrapper[4696]: I0321 08:46:30.817782 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-vgptg" Mar 21 08:46:30 crc kubenswrapper[4696]: I0321 08:46:30.817781 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-vgptg" event={"ID":"5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1","Type":"ContainerDied","Data":"1e1693dfb3f0cdea753cd28a5502921a7896a2f3488661c85f8870a8a5ccc4b7"} Mar 21 08:46:30 crc kubenswrapper[4696]: I0321 08:46:30.818060 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e1693dfb3f0cdea753cd28a5502921a7896a2f3488661c85f8870a8a5ccc4b7" Mar 21 08:46:30 crc kubenswrapper[4696]: I0321 08:46:30.819463 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"c2c63b45-de34-43a6-bb8b-aed4b00922c8","Type":"ContainerStarted","Data":"841b5ec6f560a6ef4a8391e1aeb5997eadd22513f5a63b0ab55221718769998a"} Mar 21 08:46:30 crc kubenswrapper[4696]: I0321 08:46:30.914471 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:46:30 crc kubenswrapper[4696]: I0321 08:46:30.914616 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-dmbpm" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.128750 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4vd4t-config-2tzm6"] Mar 21 08:46:31 crc kubenswrapper[4696]: E0321 08:46:31.129206 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1" containerName="mariadb-account-create-update" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.129229 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1" containerName="mariadb-account-create-update" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.129436 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1" containerName="mariadb-account-create-update" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.130230 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.133007 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.145120 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4vd4t-config-2tzm6"] Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.215365 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-var-log-ovn\") pod \"ovn-controller-4vd4t-config-2tzm6\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.215415 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-additional-scripts\") pod \"ovn-controller-4vd4t-config-2tzm6\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.215523 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-var-run\") pod \"ovn-controller-4vd4t-config-2tzm6\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.215551 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-var-run-ovn\") pod \"ovn-controller-4vd4t-config-2tzm6\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.215601 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-scripts\") pod \"ovn-controller-4vd4t-config-2tzm6\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.215638 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nblg7\" (UniqueName: \"kubernetes.io/projected/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-kube-api-access-nblg7\") pod \"ovn-controller-4vd4t-config-2tzm6\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.317653 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-var-log-ovn\") pod \"ovn-controller-4vd4t-config-2tzm6\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.317726 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-additional-scripts\") pod \"ovn-controller-4vd4t-config-2tzm6\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.317846 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-var-run\") pod \"ovn-controller-4vd4t-config-2tzm6\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.317877 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-var-run-ovn\") pod \"ovn-controller-4vd4t-config-2tzm6\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.317934 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-scripts\") pod \"ovn-controller-4vd4t-config-2tzm6\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.317959 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nblg7\" (UniqueName: \"kubernetes.io/projected/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-kube-api-access-nblg7\") pod \"ovn-controller-4vd4t-config-2tzm6\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.318016 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-var-log-ovn\") pod \"ovn-controller-4vd4t-config-2tzm6\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.318079 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-var-run\") pod \"ovn-controller-4vd4t-config-2tzm6\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.318306 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-var-run-ovn\") pod \"ovn-controller-4vd4t-config-2tzm6\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.318721 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-additional-scripts\") pod \"ovn-controller-4vd4t-config-2tzm6\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.320011 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-scripts\") pod \"ovn-controller-4vd4t-config-2tzm6\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.336452 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nblg7\" (UniqueName: \"kubernetes.io/projected/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-kube-api-access-nblg7\") pod \"ovn-controller-4vd4t-config-2tzm6\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:31 crc kubenswrapper[4696]: I0321 08:46:31.490481 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:32 crc kubenswrapper[4696]: I0321 08:46:32.532420 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-vgptg"] Mar 21 08:46:32 crc kubenswrapper[4696]: I0321 08:46:32.548758 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-vgptg"] Mar 21 08:46:32 crc kubenswrapper[4696]: I0321 08:46:32.644752 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4vd4t-config-2tzm6"] Mar 21 08:46:32 crc kubenswrapper[4696]: W0321 08:46:32.659284 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4ed4576_2ced_421a_9f8f_f6f45b6a6a36.slice/crio-117e361112527493728b1da1a565af82de22d450a58fc5427b998d587c15c8af WatchSource:0}: Error finding container 117e361112527493728b1da1a565af82de22d450a58fc5427b998d587c15c8af: Status 404 returned error can't find the container with id 117e361112527493728b1da1a565af82de22d450a58fc5427b998d587c15c8af Mar 21 08:46:32 crc kubenswrapper[4696]: I0321 08:46:32.850585 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"c2c63b45-de34-43a6-bb8b-aed4b00922c8","Type":"ContainerStarted","Data":"52e185403ef4545ad38a4d2d7f889ca2021368daa7c7dcceb25b94bacde101ff"} Mar 21 08:46:32 crc kubenswrapper[4696]: I0321 08:46:32.852466 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4vd4t-config-2tzm6" event={"ID":"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36","Type":"ContainerStarted","Data":"117e361112527493728b1da1a565af82de22d450a58fc5427b998d587c15c8af"} Mar 21 08:46:33 crc kubenswrapper[4696]: I0321 08:46:33.861511 4696 generic.go:334] "Generic (PLEG): container finished" podID="e4ed4576-2ced-421a-9f8f-f6f45b6a6a36" containerID="00f2cf65b6102c22b225f29f98f3f9ac8a2f45e936cac062139f4b8e3a6a2594" exitCode=0 Mar 21 08:46:33 crc kubenswrapper[4696]: I0321 08:46:33.861554 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4vd4t-config-2tzm6" event={"ID":"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36","Type":"ContainerDied","Data":"00f2cf65b6102c22b225f29f98f3f9ac8a2f45e936cac062139f4b8e3a6a2594"} Mar 21 08:46:34 crc kubenswrapper[4696]: I0321 08:46:34.386008 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-etc-swift\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:34 crc kubenswrapper[4696]: I0321 08:46:34.399525 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1088ee04-73e8-4aba-aad4-1cd7d26e20aa-etc-swift\") pod \"swift-storage-0\" (UID: \"1088ee04-73e8-4aba-aad4-1cd7d26e20aa\") " pod="openstack/swift-storage-0" Mar 21 08:46:34 crc kubenswrapper[4696]: I0321 08:46:34.510196 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 21 08:46:34 crc kubenswrapper[4696]: I0321 08:46:34.551470 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1" path="/var/lib/kubelet/pods/5ccfa6bc-c97b-4b79-b8cc-466cb3e676b1/volumes" Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.140500 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.272765 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.403025 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-scripts\") pod \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.403106 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-var-log-ovn\") pod \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.403168 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nblg7\" (UniqueName: \"kubernetes.io/projected/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-kube-api-access-nblg7\") pod \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.403211 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-additional-scripts\") pod \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.403215 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "e4ed4576-2ced-421a-9f8f-f6f45b6a6a36" (UID: "e4ed4576-2ced-421a-9f8f-f6f45b6a6a36"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.403228 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-var-run-ovn\") pod \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.403246 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "e4ed4576-2ced-421a-9f8f-f6f45b6a6a36" (UID: "e4ed4576-2ced-421a-9f8f-f6f45b6a6a36"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.403355 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-var-run\") pod \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\" (UID: \"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36\") " Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.403420 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-var-run" (OuterVolumeSpecName: "var-run") pod "e4ed4576-2ced-421a-9f8f-f6f45b6a6a36" (UID: "e4ed4576-2ced-421a-9f8f-f6f45b6a6a36"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.403734 4696 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.403751 4696 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.403761 4696 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-var-run\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.403787 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "e4ed4576-2ced-421a-9f8f-f6f45b6a6a36" (UID: "e4ed4576-2ced-421a-9f8f-f6f45b6a6a36"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.404047 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-scripts" (OuterVolumeSpecName: "scripts") pod "e4ed4576-2ced-421a-9f8f-f6f45b6a6a36" (UID: "e4ed4576-2ced-421a-9f8f-f6f45b6a6a36"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.427149 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-kube-api-access-nblg7" (OuterVolumeSpecName: "kube-api-access-nblg7") pod "e4ed4576-2ced-421a-9f8f-f6f45b6a6a36" (UID: "e4ed4576-2ced-421a-9f8f-f6f45b6a6a36"). InnerVolumeSpecName "kube-api-access-nblg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.505744 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.505775 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nblg7\" (UniqueName: \"kubernetes.io/projected/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-kube-api-access-nblg7\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.505785 4696 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.880709 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4vd4t-config-2tzm6" event={"ID":"e4ed4576-2ced-421a-9f8f-f6f45b6a6a36","Type":"ContainerDied","Data":"117e361112527493728b1da1a565af82de22d450a58fc5427b998d587c15c8af"} Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.881037 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="117e361112527493728b1da1a565af82de22d450a58fc5427b998d587c15c8af" Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.880924 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4vd4t-config-2tzm6" Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.882316 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1088ee04-73e8-4aba-aad4-1cd7d26e20aa","Type":"ContainerStarted","Data":"2d42ecd13fcdd02745468f7808990b716478d61df5071f4e284aee5b6bd70dec"} Mar 21 08:46:35 crc kubenswrapper[4696]: I0321 08:46:35.933754 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-4vd4t" Mar 21 08:46:36 crc kubenswrapper[4696]: I0321 08:46:36.090972 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:46:36 crc kubenswrapper[4696]: I0321 08:46:36.256555 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-b2dkt"] Mar 21 08:46:36 crc kubenswrapper[4696]: E0321 08:46:36.256923 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4ed4576-2ced-421a-9f8f-f6f45b6a6a36" containerName="ovn-config" Mar 21 08:46:36 crc kubenswrapper[4696]: I0321 08:46:36.256936 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4ed4576-2ced-421a-9f8f-f6f45b6a6a36" containerName="ovn-config" Mar 21 08:46:36 crc kubenswrapper[4696]: I0321 08:46:36.257137 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4ed4576-2ced-421a-9f8f-f6f45b6a6a36" containerName="ovn-config" Mar 21 08:46:36 crc kubenswrapper[4696]: I0321 08:46:36.259016 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-b2dkt" Mar 21 08:46:36 crc kubenswrapper[4696]: I0321 08:46:36.262841 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 21 08:46:36 crc kubenswrapper[4696]: I0321 08:46:36.267298 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-b2dkt"] Mar 21 08:46:36 crc kubenswrapper[4696]: I0321 08:46:36.392987 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 21 08:46:36 crc kubenswrapper[4696]: I0321 08:46:36.405931 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-4vd4t-config-2tzm6"] Mar 21 08:46:36 crc kubenswrapper[4696]: I0321 08:46:36.417219 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-4vd4t-config-2tzm6"] Mar 21 08:46:36 crc kubenswrapper[4696]: I0321 08:46:36.427958 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54ede1d0-752d-4967-a158-d41b7d56ddba-operator-scripts\") pod \"root-account-create-update-b2dkt\" (UID: \"54ede1d0-752d-4967-a158-d41b7d56ddba\") " pod="openstack/root-account-create-update-b2dkt" Mar 21 08:46:36 crc kubenswrapper[4696]: I0321 08:46:36.428099 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5d7m\" (UniqueName: \"kubernetes.io/projected/54ede1d0-752d-4967-a158-d41b7d56ddba-kube-api-access-x5d7m\") pod \"root-account-create-update-b2dkt\" (UID: \"54ede1d0-752d-4967-a158-d41b7d56ddba\") " pod="openstack/root-account-create-update-b2dkt" Mar 21 08:46:36 crc kubenswrapper[4696]: I0321 08:46:36.529545 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5d7m\" (UniqueName: \"kubernetes.io/projected/54ede1d0-752d-4967-a158-d41b7d56ddba-kube-api-access-x5d7m\") pod \"root-account-create-update-b2dkt\" (UID: \"54ede1d0-752d-4967-a158-d41b7d56ddba\") " pod="openstack/root-account-create-update-b2dkt" Mar 21 08:46:36 crc kubenswrapper[4696]: I0321 08:46:36.529901 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54ede1d0-752d-4967-a158-d41b7d56ddba-operator-scripts\") pod \"root-account-create-update-b2dkt\" (UID: \"54ede1d0-752d-4967-a158-d41b7d56ddba\") " pod="openstack/root-account-create-update-b2dkt" Mar 21 08:46:36 crc kubenswrapper[4696]: I0321 08:46:36.531025 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54ede1d0-752d-4967-a158-d41b7d56ddba-operator-scripts\") pod \"root-account-create-update-b2dkt\" (UID: \"54ede1d0-752d-4967-a158-d41b7d56ddba\") " pod="openstack/root-account-create-update-b2dkt" Mar 21 08:46:36 crc kubenswrapper[4696]: I0321 08:46:36.546777 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4ed4576-2ced-421a-9f8f-f6f45b6a6a36" path="/var/lib/kubelet/pods/e4ed4576-2ced-421a-9f8f-f6f45b6a6a36/volumes" Mar 21 08:46:36 crc kubenswrapper[4696]: I0321 08:46:36.563540 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5d7m\" (UniqueName: \"kubernetes.io/projected/54ede1d0-752d-4967-a158-d41b7d56ddba-kube-api-access-x5d7m\") pod \"root-account-create-update-b2dkt\" (UID: \"54ede1d0-752d-4967-a158-d41b7d56ddba\") " pod="openstack/root-account-create-update-b2dkt" Mar 21 08:46:36 crc kubenswrapper[4696]: I0321 08:46:36.581620 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-b2dkt" Mar 21 08:46:37 crc kubenswrapper[4696]: I0321 08:46:37.212748 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-b2dkt"] Mar 21 08:46:37 crc kubenswrapper[4696]: W0321 08:46:37.222165 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54ede1d0_752d_4967_a158_d41b7d56ddba.slice/crio-dc821e08b634b2f9bea2956ee7fa96ae605788a3abcfa6ee85c1d7ec3e1ac768 WatchSource:0}: Error finding container dc821e08b634b2f9bea2956ee7fa96ae605788a3abcfa6ee85c1d7ec3e1ac768: Status 404 returned error can't find the container with id dc821e08b634b2f9bea2956ee7fa96ae605788a3abcfa6ee85c1d7ec3e1ac768 Mar 21 08:46:37 crc kubenswrapper[4696]: I0321 08:46:37.931854 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-b2dkt" event={"ID":"54ede1d0-752d-4967-a158-d41b7d56ddba","Type":"ContainerStarted","Data":"dbc48018407d29c23bb500596ff5dfd1c996e0eee22ffe4e1a86f17995f71a41"} Mar 21 08:46:37 crc kubenswrapper[4696]: I0321 08:46:37.932211 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-b2dkt" event={"ID":"54ede1d0-752d-4967-a158-d41b7d56ddba","Type":"ContainerStarted","Data":"dc821e08b634b2f9bea2956ee7fa96ae605788a3abcfa6ee85c1d7ec3e1ac768"} Mar 21 08:46:37 crc kubenswrapper[4696]: I0321 08:46:37.943541 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1088ee04-73e8-4aba-aad4-1cd7d26e20aa","Type":"ContainerStarted","Data":"27b9ebbe9a99d3befd227634480cf7ab4b3922d89d3199831aa20f648acead22"} Mar 21 08:46:37 crc kubenswrapper[4696]: I0321 08:46:37.943589 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1088ee04-73e8-4aba-aad4-1cd7d26e20aa","Type":"ContainerStarted","Data":"db2cbb719d5b2b7ffd976c38b4c33efd0dd4bb6fb0fb298a95fe056f794da85f"} Mar 21 08:46:37 crc kubenswrapper[4696]: I0321 08:46:37.943603 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1088ee04-73e8-4aba-aad4-1cd7d26e20aa","Type":"ContainerStarted","Data":"f1de91efcf40dac15e0528f8760a8e83a41ac41b3996097ade4bdeb452c9957a"} Mar 21 08:46:37 crc kubenswrapper[4696]: I0321 08:46:37.946078 4696 generic.go:334] "Generic (PLEG): container finished" podID="c2c63b45-de34-43a6-bb8b-aed4b00922c8" containerID="52e185403ef4545ad38a4d2d7f889ca2021368daa7c7dcceb25b94bacde101ff" exitCode=0 Mar 21 08:46:37 crc kubenswrapper[4696]: I0321 08:46:37.946110 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"c2c63b45-de34-43a6-bb8b-aed4b00922c8","Type":"ContainerDied","Data":"52e185403ef4545ad38a4d2d7f889ca2021368daa7c7dcceb25b94bacde101ff"} Mar 21 08:46:37 crc kubenswrapper[4696]: I0321 08:46:37.962071 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-b2dkt" podStartSLOduration=1.962048168 podStartE2EDuration="1.962048168s" podCreationTimestamp="2026-03-21 08:46:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:46:37.950574347 +0000 UTC m=+1132.071455050" watchObservedRunningTime="2026-03-21 08:46:37.962048168 +0000 UTC m=+1132.082928881" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.185955 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-mtbts"] Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.192387 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-mtbts" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.195560 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-mtbts"] Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.273601 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5be1b1b8-ee1e-47cd-afbe-e1a96cee697a-operator-scripts\") pod \"cinder-db-create-mtbts\" (UID: \"5be1b1b8-ee1e-47cd-afbe-e1a96cee697a\") " pod="openstack/cinder-db-create-mtbts" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.273710 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmmjg\" (UniqueName: \"kubernetes.io/projected/5be1b1b8-ee1e-47cd-afbe-e1a96cee697a-kube-api-access-cmmjg\") pod \"cinder-db-create-mtbts\" (UID: \"5be1b1b8-ee1e-47cd-afbe-e1a96cee697a\") " pod="openstack/cinder-db-create-mtbts" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.320249 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-f359-account-create-update-2jqvh"] Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.321534 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f359-account-create-update-2jqvh" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.325070 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.332409 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f359-account-create-update-2jqvh"] Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.375125 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmmjg\" (UniqueName: \"kubernetes.io/projected/5be1b1b8-ee1e-47cd-afbe-e1a96cee697a-kube-api-access-cmmjg\") pod \"cinder-db-create-mtbts\" (UID: \"5be1b1b8-ee1e-47cd-afbe-e1a96cee697a\") " pod="openstack/cinder-db-create-mtbts" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.375196 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec226d0f-1ee8-481b-85a5-280fe2d17810-operator-scripts\") pod \"cinder-f359-account-create-update-2jqvh\" (UID: \"ec226d0f-1ee8-481b-85a5-280fe2d17810\") " pod="openstack/cinder-f359-account-create-update-2jqvh" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.375312 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5be1b1b8-ee1e-47cd-afbe-e1a96cee697a-operator-scripts\") pod \"cinder-db-create-mtbts\" (UID: \"5be1b1b8-ee1e-47cd-afbe-e1a96cee697a\") " pod="openstack/cinder-db-create-mtbts" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.375371 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6d5n\" (UniqueName: \"kubernetes.io/projected/ec226d0f-1ee8-481b-85a5-280fe2d17810-kube-api-access-j6d5n\") pod \"cinder-f359-account-create-update-2jqvh\" (UID: \"ec226d0f-1ee8-481b-85a5-280fe2d17810\") " pod="openstack/cinder-f359-account-create-update-2jqvh" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.376053 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5be1b1b8-ee1e-47cd-afbe-e1a96cee697a-operator-scripts\") pod \"cinder-db-create-mtbts\" (UID: \"5be1b1b8-ee1e-47cd-afbe-e1a96cee697a\") " pod="openstack/cinder-db-create-mtbts" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.398327 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmmjg\" (UniqueName: \"kubernetes.io/projected/5be1b1b8-ee1e-47cd-afbe-e1a96cee697a-kube-api-access-cmmjg\") pod \"cinder-db-create-mtbts\" (UID: \"5be1b1b8-ee1e-47cd-afbe-e1a96cee697a\") " pod="openstack/cinder-db-create-mtbts" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.477332 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6d5n\" (UniqueName: \"kubernetes.io/projected/ec226d0f-1ee8-481b-85a5-280fe2d17810-kube-api-access-j6d5n\") pod \"cinder-f359-account-create-update-2jqvh\" (UID: \"ec226d0f-1ee8-481b-85a5-280fe2d17810\") " pod="openstack/cinder-f359-account-create-update-2jqvh" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.477436 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec226d0f-1ee8-481b-85a5-280fe2d17810-operator-scripts\") pod \"cinder-f359-account-create-update-2jqvh\" (UID: \"ec226d0f-1ee8-481b-85a5-280fe2d17810\") " pod="openstack/cinder-f359-account-create-update-2jqvh" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.478179 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec226d0f-1ee8-481b-85a5-280fe2d17810-operator-scripts\") pod \"cinder-f359-account-create-update-2jqvh\" (UID: \"ec226d0f-1ee8-481b-85a5-280fe2d17810\") " pod="openstack/cinder-f359-account-create-update-2jqvh" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.486912 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-db-create-tjmmf"] Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.499389 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-create-tjmmf"] Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.501057 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-tjmmf" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.513250 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-mtbts" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.513571 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-2ef2-account-create-update-9pxpg"] Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.514752 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-2ef2-account-create-update-9pxpg" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.521709 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-db-secret" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.525149 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6d5n\" (UniqueName: \"kubernetes.io/projected/ec226d0f-1ee8-481b-85a5-280fe2d17810-kube-api-access-j6d5n\") pod \"cinder-f359-account-create-update-2jqvh\" (UID: \"ec226d0f-1ee8-481b-85a5-280fe2d17810\") " pod="openstack/cinder-f359-account-create-update-2jqvh" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.573384 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-2ef2-account-create-update-9pxpg"] Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.581039 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr8wq\" (UniqueName: \"kubernetes.io/projected/cf260a42-d042-4130-bc96-9e12dcb253c4-kube-api-access-tr8wq\") pod \"cloudkitty-db-create-tjmmf\" (UID: \"cf260a42-d042-4130-bc96-9e12dcb253c4\") " pod="openstack/cloudkitty-db-create-tjmmf" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.581148 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vklxz\" (UniqueName: \"kubernetes.io/projected/17520989-db17-4ad3-8324-279698af93e3-kube-api-access-vklxz\") pod \"cloudkitty-2ef2-account-create-update-9pxpg\" (UID: \"17520989-db17-4ad3-8324-279698af93e3\") " pod="openstack/cloudkitty-2ef2-account-create-update-9pxpg" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.581176 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17520989-db17-4ad3-8324-279698af93e3-operator-scripts\") pod \"cloudkitty-2ef2-account-create-update-9pxpg\" (UID: \"17520989-db17-4ad3-8324-279698af93e3\") " pod="openstack/cloudkitty-2ef2-account-create-update-9pxpg" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.581224 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf260a42-d042-4130-bc96-9e12dcb253c4-operator-scripts\") pod \"cloudkitty-db-create-tjmmf\" (UID: \"cf260a42-d042-4130-bc96-9e12dcb253c4\") " pod="openstack/cloudkitty-db-create-tjmmf" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.627702 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-rff8n"] Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.629221 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rff8n" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.649727 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rff8n"] Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.699288 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr8wq\" (UniqueName: \"kubernetes.io/projected/cf260a42-d042-4130-bc96-9e12dcb253c4-kube-api-access-tr8wq\") pod \"cloudkitty-db-create-tjmmf\" (UID: \"cf260a42-d042-4130-bc96-9e12dcb253c4\") " pod="openstack/cloudkitty-db-create-tjmmf" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.699508 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnfxq\" (UniqueName: \"kubernetes.io/projected/dc7fc65b-70b4-46cd-84e5-16285d93f295-kube-api-access-mnfxq\") pod \"barbican-db-create-rff8n\" (UID: \"dc7fc65b-70b4-46cd-84e5-16285d93f295\") " pod="openstack/barbican-db-create-rff8n" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.743468 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc7fc65b-70b4-46cd-84e5-16285d93f295-operator-scripts\") pod \"barbican-db-create-rff8n\" (UID: \"dc7fc65b-70b4-46cd-84e5-16285d93f295\") " pod="openstack/barbican-db-create-rff8n" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.743510 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vklxz\" (UniqueName: \"kubernetes.io/projected/17520989-db17-4ad3-8324-279698af93e3-kube-api-access-vklxz\") pod \"cloudkitty-2ef2-account-create-update-9pxpg\" (UID: \"17520989-db17-4ad3-8324-279698af93e3\") " pod="openstack/cloudkitty-2ef2-account-create-update-9pxpg" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.743566 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17520989-db17-4ad3-8324-279698af93e3-operator-scripts\") pod \"cloudkitty-2ef2-account-create-update-9pxpg\" (UID: \"17520989-db17-4ad3-8324-279698af93e3\") " pod="openstack/cloudkitty-2ef2-account-create-update-9pxpg" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.743630 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf260a42-d042-4130-bc96-9e12dcb253c4-operator-scripts\") pod \"cloudkitty-db-create-tjmmf\" (UID: \"cf260a42-d042-4130-bc96-9e12dcb253c4\") " pod="openstack/cloudkitty-db-create-tjmmf" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.735613 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f359-account-create-update-2jqvh" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.744707 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf260a42-d042-4130-bc96-9e12dcb253c4-operator-scripts\") pod \"cloudkitty-db-create-tjmmf\" (UID: \"cf260a42-d042-4130-bc96-9e12dcb253c4\") " pod="openstack/cloudkitty-db-create-tjmmf" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.745633 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17520989-db17-4ad3-8324-279698af93e3-operator-scripts\") pod \"cloudkitty-2ef2-account-create-update-9pxpg\" (UID: \"17520989-db17-4ad3-8324-279698af93e3\") " pod="openstack/cloudkitty-2ef2-account-create-update-9pxpg" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.739347 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-4098-account-create-update-mn786"] Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.754891 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4098-account-create-update-mn786" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.758634 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr8wq\" (UniqueName: \"kubernetes.io/projected/cf260a42-d042-4130-bc96-9e12dcb253c4-kube-api-access-tr8wq\") pod \"cloudkitty-db-create-tjmmf\" (UID: \"cf260a42-d042-4130-bc96-9e12dcb253c4\") " pod="openstack/cloudkitty-db-create-tjmmf" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.760438 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.793612 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vklxz\" (UniqueName: \"kubernetes.io/projected/17520989-db17-4ad3-8324-279698af93e3-kube-api-access-vklxz\") pod \"cloudkitty-2ef2-account-create-update-9pxpg\" (UID: \"17520989-db17-4ad3-8324-279698af93e3\") " pod="openstack/cloudkitty-2ef2-account-create-update-9pxpg" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.807075 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-4098-account-create-update-mn786"] Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.821381 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-tjmmf" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.836427 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-2ef2-account-create-update-9pxpg" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.840154 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-trn7c"] Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.841392 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-trn7c" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.844985 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.845178 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-s9frx" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.845282 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.845377 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.856000 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-trn7c"] Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.870195 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-52e3-account-create-update-5pmnh"] Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.871352 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-52e3-account-create-update-5pmnh" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.881595 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.887879 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-52e3-account-create-update-5pmnh"] Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.894082 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnfxq\" (UniqueName: \"kubernetes.io/projected/dc7fc65b-70b4-46cd-84e5-16285d93f295-kube-api-access-mnfxq\") pod \"barbican-db-create-rff8n\" (UID: \"dc7fc65b-70b4-46cd-84e5-16285d93f295\") " pod="openstack/barbican-db-create-rff8n" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.894128 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc7fc65b-70b4-46cd-84e5-16285d93f295-operator-scripts\") pod \"barbican-db-create-rff8n\" (UID: \"dc7fc65b-70b4-46cd-84e5-16285d93f295\") " pod="openstack/barbican-db-create-rff8n" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.894156 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a6217ca-7143-41fd-ae0d-f96bde35029c-operator-scripts\") pod \"barbican-4098-account-create-update-mn786\" (UID: \"1a6217ca-7143-41fd-ae0d-f96bde35029c\") " pod="openstack/barbican-4098-account-create-update-mn786" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.894181 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mn44v\" (UniqueName: \"kubernetes.io/projected/1a6217ca-7143-41fd-ae0d-f96bde35029c-kube-api-access-mn44v\") pod \"barbican-4098-account-create-update-mn786\" (UID: \"1a6217ca-7143-41fd-ae0d-f96bde35029c\") " pod="openstack/barbican-4098-account-create-update-mn786" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.895442 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc7fc65b-70b4-46cd-84e5-16285d93f295-operator-scripts\") pod \"barbican-db-create-rff8n\" (UID: \"dc7fc65b-70b4-46cd-84e5-16285d93f295\") " pod="openstack/barbican-db-create-rff8n" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.918708 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-jc8bj"] Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.920024 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jc8bj" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.926625 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-jc8bj"] Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.931837 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnfxq\" (UniqueName: \"kubernetes.io/projected/dc7fc65b-70b4-46cd-84e5-16285d93f295-kube-api-access-mnfxq\") pod \"barbican-db-create-rff8n\" (UID: \"dc7fc65b-70b4-46cd-84e5-16285d93f295\") " pod="openstack/barbican-db-create-rff8n" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.953508 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rff8n" Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.975840 4696 generic.go:334] "Generic (PLEG): container finished" podID="54ede1d0-752d-4967-a158-d41b7d56ddba" containerID="dbc48018407d29c23bb500596ff5dfd1c996e0eee22ffe4e1a86f17995f71a41" exitCode=0 Mar 21 08:46:38 crc kubenswrapper[4696]: I0321 08:46:38.975927 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-b2dkt" event={"ID":"54ede1d0-752d-4967-a158-d41b7d56ddba","Type":"ContainerDied","Data":"dbc48018407d29c23bb500596ff5dfd1c996e0eee22ffe4e1a86f17995f71a41"} Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.011081 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92718cf1-7a8f-42cd-aadc-df800f2499d4-combined-ca-bundle\") pod \"keystone-db-sync-trn7c\" (UID: \"92718cf1-7a8f-42cd-aadc-df800f2499d4\") " pod="openstack/keystone-db-sync-trn7c" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.011126 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92718cf1-7a8f-42cd-aadc-df800f2499d4-config-data\") pod \"keystone-db-sync-trn7c\" (UID: \"92718cf1-7a8f-42cd-aadc-df800f2499d4\") " pod="openstack/keystone-db-sync-trn7c" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.011149 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bac528ca-04da-4a92-9b88-b9a3d226cb94-operator-scripts\") pod \"neutron-52e3-account-create-update-5pmnh\" (UID: \"bac528ca-04da-4a92-9b88-b9a3d226cb94\") " pod="openstack/neutron-52e3-account-create-update-5pmnh" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.011181 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kppb\" (UniqueName: \"kubernetes.io/projected/bac528ca-04da-4a92-9b88-b9a3d226cb94-kube-api-access-9kppb\") pod \"neutron-52e3-account-create-update-5pmnh\" (UID: \"bac528ca-04da-4a92-9b88-b9a3d226cb94\") " pod="openstack/neutron-52e3-account-create-update-5pmnh" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.011256 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb455\" (UniqueName: \"kubernetes.io/projected/80468ce8-5377-4423-9e04-b58d1d0910d7-kube-api-access-wb455\") pod \"neutron-db-create-jc8bj\" (UID: \"80468ce8-5377-4423-9e04-b58d1d0910d7\") " pod="openstack/neutron-db-create-jc8bj" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.011282 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a6217ca-7143-41fd-ae0d-f96bde35029c-operator-scripts\") pod \"barbican-4098-account-create-update-mn786\" (UID: \"1a6217ca-7143-41fd-ae0d-f96bde35029c\") " pod="openstack/barbican-4098-account-create-update-mn786" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.011313 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mn44v\" (UniqueName: \"kubernetes.io/projected/1a6217ca-7143-41fd-ae0d-f96bde35029c-kube-api-access-mn44v\") pod \"barbican-4098-account-create-update-mn786\" (UID: \"1a6217ca-7143-41fd-ae0d-f96bde35029c\") " pod="openstack/barbican-4098-account-create-update-mn786" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.011362 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80468ce8-5377-4423-9e04-b58d1d0910d7-operator-scripts\") pod \"neutron-db-create-jc8bj\" (UID: \"80468ce8-5377-4423-9e04-b58d1d0910d7\") " pod="openstack/neutron-db-create-jc8bj" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.011389 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdh62\" (UniqueName: \"kubernetes.io/projected/92718cf1-7a8f-42cd-aadc-df800f2499d4-kube-api-access-bdh62\") pod \"keystone-db-sync-trn7c\" (UID: \"92718cf1-7a8f-42cd-aadc-df800f2499d4\") " pod="openstack/keystone-db-sync-trn7c" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.012118 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a6217ca-7143-41fd-ae0d-f96bde35029c-operator-scripts\") pod \"barbican-4098-account-create-update-mn786\" (UID: \"1a6217ca-7143-41fd-ae0d-f96bde35029c\") " pod="openstack/barbican-4098-account-create-update-mn786" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.023983 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1088ee04-73e8-4aba-aad4-1cd7d26e20aa","Type":"ContainerStarted","Data":"8b88f869255ddefa1b9b5d46b313b178838ec3d062966a58297959c87e374488"} Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.038549 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"c2c63b45-de34-43a6-bb8b-aed4b00922c8","Type":"ContainerStarted","Data":"10300a08b2387f13b5bd7c7b123c2b1f2fe970a5973eacceb607c3a768fe46fb"} Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.112874 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80468ce8-5377-4423-9e04-b58d1d0910d7-operator-scripts\") pod \"neutron-db-create-jc8bj\" (UID: \"80468ce8-5377-4423-9e04-b58d1d0910d7\") " pod="openstack/neutron-db-create-jc8bj" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.113270 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdh62\" (UniqueName: \"kubernetes.io/projected/92718cf1-7a8f-42cd-aadc-df800f2499d4-kube-api-access-bdh62\") pod \"keystone-db-sync-trn7c\" (UID: \"92718cf1-7a8f-42cd-aadc-df800f2499d4\") " pod="openstack/keystone-db-sync-trn7c" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.113335 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92718cf1-7a8f-42cd-aadc-df800f2499d4-combined-ca-bundle\") pod \"keystone-db-sync-trn7c\" (UID: \"92718cf1-7a8f-42cd-aadc-df800f2499d4\") " pod="openstack/keystone-db-sync-trn7c" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.117619 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92718cf1-7a8f-42cd-aadc-df800f2499d4-config-data\") pod \"keystone-db-sync-trn7c\" (UID: \"92718cf1-7a8f-42cd-aadc-df800f2499d4\") " pod="openstack/keystone-db-sync-trn7c" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.117797 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bac528ca-04da-4a92-9b88-b9a3d226cb94-operator-scripts\") pod \"neutron-52e3-account-create-update-5pmnh\" (UID: \"bac528ca-04da-4a92-9b88-b9a3d226cb94\") " pod="openstack/neutron-52e3-account-create-update-5pmnh" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.118136 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kppb\" (UniqueName: \"kubernetes.io/projected/bac528ca-04da-4a92-9b88-b9a3d226cb94-kube-api-access-9kppb\") pod \"neutron-52e3-account-create-update-5pmnh\" (UID: \"bac528ca-04da-4a92-9b88-b9a3d226cb94\") " pod="openstack/neutron-52e3-account-create-update-5pmnh" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.118679 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb455\" (UniqueName: \"kubernetes.io/projected/80468ce8-5377-4423-9e04-b58d1d0910d7-kube-api-access-wb455\") pod \"neutron-db-create-jc8bj\" (UID: \"80468ce8-5377-4423-9e04-b58d1d0910d7\") " pod="openstack/neutron-db-create-jc8bj" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.121541 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bac528ca-04da-4a92-9b88-b9a3d226cb94-operator-scripts\") pod \"neutron-52e3-account-create-update-5pmnh\" (UID: \"bac528ca-04da-4a92-9b88-b9a3d226cb94\") " pod="openstack/neutron-52e3-account-create-update-5pmnh" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.129911 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92718cf1-7a8f-42cd-aadc-df800f2499d4-config-data\") pod \"keystone-db-sync-trn7c\" (UID: \"92718cf1-7a8f-42cd-aadc-df800f2499d4\") " pod="openstack/keystone-db-sync-trn7c" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.130412 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92718cf1-7a8f-42cd-aadc-df800f2499d4-combined-ca-bundle\") pod \"keystone-db-sync-trn7c\" (UID: \"92718cf1-7a8f-42cd-aadc-df800f2499d4\") " pod="openstack/keystone-db-sync-trn7c" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.144273 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kppb\" (UniqueName: \"kubernetes.io/projected/bac528ca-04da-4a92-9b88-b9a3d226cb94-kube-api-access-9kppb\") pod \"neutron-52e3-account-create-update-5pmnh\" (UID: \"bac528ca-04da-4a92-9b88-b9a3d226cb94\") " pod="openstack/neutron-52e3-account-create-update-5pmnh" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.154549 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdh62\" (UniqueName: \"kubernetes.io/projected/92718cf1-7a8f-42cd-aadc-df800f2499d4-kube-api-access-bdh62\") pod \"keystone-db-sync-trn7c\" (UID: \"92718cf1-7a8f-42cd-aadc-df800f2499d4\") " pod="openstack/keystone-db-sync-trn7c" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.168509 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mn44v\" (UniqueName: \"kubernetes.io/projected/1a6217ca-7143-41fd-ae0d-f96bde35029c-kube-api-access-mn44v\") pod \"barbican-4098-account-create-update-mn786\" (UID: \"1a6217ca-7143-41fd-ae0d-f96bde35029c\") " pod="openstack/barbican-4098-account-create-update-mn786" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.169488 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80468ce8-5377-4423-9e04-b58d1d0910d7-operator-scripts\") pod \"neutron-db-create-jc8bj\" (UID: \"80468ce8-5377-4423-9e04-b58d1d0910d7\") " pod="openstack/neutron-db-create-jc8bj" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.172580 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb455\" (UniqueName: \"kubernetes.io/projected/80468ce8-5377-4423-9e04-b58d1d0910d7-kube-api-access-wb455\") pod \"neutron-db-create-jc8bj\" (UID: \"80468ce8-5377-4423-9e04-b58d1d0910d7\") " pod="openstack/neutron-db-create-jc8bj" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.219208 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4098-account-create-update-mn786" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.257313 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-trn7c" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.281450 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-52e3-account-create-update-5pmnh" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.282872 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-mtbts"] Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.301103 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jc8bj" Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.316665 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f359-account-create-update-2jqvh"] Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.583183 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-2ef2-account-create-update-9pxpg"] Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.630517 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-create-tjmmf"] Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.778464 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rff8n"] Mar 21 08:46:39 crc kubenswrapper[4696]: W0321 08:46:39.898013 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc7fc65b_70b4_46cd_84e5_16285d93f295.slice/crio-c63dee5f82e8064bc80b8774e641cfdff9efe69ef0ac4f5b4ff5e03b266fe549 WatchSource:0}: Error finding container c63dee5f82e8064bc80b8774e641cfdff9efe69ef0ac4f5b4ff5e03b266fe549: Status 404 returned error can't find the container with id c63dee5f82e8064bc80b8774e641cfdff9efe69ef0ac4f5b4ff5e03b266fe549 Mar 21 08:46:39 crc kubenswrapper[4696]: I0321 08:46:39.925707 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-4098-account-create-update-mn786"] Mar 21 08:46:40 crc kubenswrapper[4696]: I0321 08:46:40.011878 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-52e3-account-create-update-5pmnh"] Mar 21 08:46:40 crc kubenswrapper[4696]: I0321 08:46:40.017703 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-trn7c"] Mar 21 08:46:40 crc kubenswrapper[4696]: I0321 08:46:40.083508 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-tjmmf" event={"ID":"cf260a42-d042-4130-bc96-9e12dcb253c4","Type":"ContainerStarted","Data":"da3e06e5c2261fd2a3d9b4990569f2bf8c3f9ac0bca2060fee118dd083e2b3d3"} Mar 21 08:46:40 crc kubenswrapper[4696]: I0321 08:46:40.084718 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-jc8bj"] Mar 21 08:46:40 crc kubenswrapper[4696]: I0321 08:46:40.087558 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f359-account-create-update-2jqvh" event={"ID":"ec226d0f-1ee8-481b-85a5-280fe2d17810","Type":"ContainerStarted","Data":"ff845693c915bd2420f675864f12930073f36bb1cfafce4dc20bad658632af06"} Mar 21 08:46:40 crc kubenswrapper[4696]: I0321 08:46:40.087602 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f359-account-create-update-2jqvh" event={"ID":"ec226d0f-1ee8-481b-85a5-280fe2d17810","Type":"ContainerStarted","Data":"70db639e7e34186303d1a788ca83dbcce42d34c2d2454c9312209ba35666a955"} Mar 21 08:46:40 crc kubenswrapper[4696]: I0321 08:46:40.089157 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-2ef2-account-create-update-9pxpg" event={"ID":"17520989-db17-4ad3-8324-279698af93e3","Type":"ContainerStarted","Data":"ca2886eb12cee6010f9f5dddb018051353dae74350b01dfee2b1ced498bdbbcd"} Mar 21 08:46:40 crc kubenswrapper[4696]: I0321 08:46:40.110470 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-mtbts" event={"ID":"5be1b1b8-ee1e-47cd-afbe-e1a96cee697a","Type":"ContainerStarted","Data":"f4a25a7aa309127c8da648696529dc7f5675491cd67081147bf74654f558c98b"} Mar 21 08:46:40 crc kubenswrapper[4696]: I0321 08:46:40.110519 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-mtbts" event={"ID":"5be1b1b8-ee1e-47cd-afbe-e1a96cee697a","Type":"ContainerStarted","Data":"205add55b36fb047113d7717db8ce5a124cf5adadf9e2fc023fcd7ff0c89c3fa"} Mar 21 08:46:40 crc kubenswrapper[4696]: I0321 08:46:40.113956 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rff8n" event={"ID":"dc7fc65b-70b4-46cd-84e5-16285d93f295","Type":"ContainerStarted","Data":"c63dee5f82e8064bc80b8774e641cfdff9efe69ef0ac4f5b4ff5e03b266fe549"} Mar 21 08:46:40 crc kubenswrapper[4696]: I0321 08:46:40.151557 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-mtbts" podStartSLOduration=2.151540626 podStartE2EDuration="2.151540626s" podCreationTimestamp="2026-03-21 08:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:46:40.149189621 +0000 UTC m=+1134.270070334" watchObservedRunningTime="2026-03-21 08:46:40.151540626 +0000 UTC m=+1134.272421339" Mar 21 08:46:40 crc kubenswrapper[4696]: I0321 08:46:40.155319 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-f359-account-create-update-2jqvh" podStartSLOduration=2.155309891 podStartE2EDuration="2.155309891s" podCreationTimestamp="2026-03-21 08:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:46:40.123031131 +0000 UTC m=+1134.243911844" watchObservedRunningTime="2026-03-21 08:46:40.155309891 +0000 UTC m=+1134.276190604" Mar 21 08:46:40 crc kubenswrapper[4696]: I0321 08:46:40.657504 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="d6518eba-34ec-4a38-b279-cd8bb11e7e24" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 21 08:46:41 crc kubenswrapper[4696]: I0321 08:46:41.122765 4696 generic.go:334] "Generic (PLEG): container finished" podID="ec226d0f-1ee8-481b-85a5-280fe2d17810" containerID="ff845693c915bd2420f675864f12930073f36bb1cfafce4dc20bad658632af06" exitCode=0 Mar 21 08:46:41 crc kubenswrapper[4696]: I0321 08:46:41.122899 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f359-account-create-update-2jqvh" event={"ID":"ec226d0f-1ee8-481b-85a5-280fe2d17810","Type":"ContainerDied","Data":"ff845693c915bd2420f675864f12930073f36bb1cfafce4dc20bad658632af06"} Mar 21 08:46:41 crc kubenswrapper[4696]: I0321 08:46:41.124617 4696 generic.go:334] "Generic (PLEG): container finished" podID="17520989-db17-4ad3-8324-279698af93e3" containerID="2fbc10bfc4d089694e602949ff8f54651e8ede44ee5070f6939028e1a47a3406" exitCode=0 Mar 21 08:46:41 crc kubenswrapper[4696]: I0321 08:46:41.124676 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-2ef2-account-create-update-9pxpg" event={"ID":"17520989-db17-4ad3-8324-279698af93e3","Type":"ContainerDied","Data":"2fbc10bfc4d089694e602949ff8f54651e8ede44ee5070f6939028e1a47a3406"} Mar 21 08:46:41 crc kubenswrapper[4696]: I0321 08:46:41.126891 4696 generic.go:334] "Generic (PLEG): container finished" podID="5be1b1b8-ee1e-47cd-afbe-e1a96cee697a" containerID="f4a25a7aa309127c8da648696529dc7f5675491cd67081147bf74654f558c98b" exitCode=0 Mar 21 08:46:41 crc kubenswrapper[4696]: I0321 08:46:41.126949 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-mtbts" event={"ID":"5be1b1b8-ee1e-47cd-afbe-e1a96cee697a","Type":"ContainerDied","Data":"f4a25a7aa309127c8da648696529dc7f5675491cd67081147bf74654f558c98b"} Mar 21 08:46:41 crc kubenswrapper[4696]: I0321 08:46:41.128748 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"c2c63b45-de34-43a6-bb8b-aed4b00922c8","Type":"ContainerStarted","Data":"c5eed5bcb42135cc5451fb8a474d31a04efc9890006cfb44645e9923e2a78391"} Mar 21 08:46:47 crc kubenswrapper[4696]: I0321 08:46:47.568947 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-b2dkt" Mar 21 08:46:47 crc kubenswrapper[4696]: I0321 08:46:47.694989 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5d7m\" (UniqueName: \"kubernetes.io/projected/54ede1d0-752d-4967-a158-d41b7d56ddba-kube-api-access-x5d7m\") pod \"54ede1d0-752d-4967-a158-d41b7d56ddba\" (UID: \"54ede1d0-752d-4967-a158-d41b7d56ddba\") " Mar 21 08:46:47 crc kubenswrapper[4696]: I0321 08:46:47.695228 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54ede1d0-752d-4967-a158-d41b7d56ddba-operator-scripts\") pod \"54ede1d0-752d-4967-a158-d41b7d56ddba\" (UID: \"54ede1d0-752d-4967-a158-d41b7d56ddba\") " Mar 21 08:46:47 crc kubenswrapper[4696]: I0321 08:46:47.695601 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54ede1d0-752d-4967-a158-d41b7d56ddba-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "54ede1d0-752d-4967-a158-d41b7d56ddba" (UID: "54ede1d0-752d-4967-a158-d41b7d56ddba"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:47 crc kubenswrapper[4696]: I0321 08:46:47.701097 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54ede1d0-752d-4967-a158-d41b7d56ddba-kube-api-access-x5d7m" (OuterVolumeSpecName: "kube-api-access-x5d7m") pod "54ede1d0-752d-4967-a158-d41b7d56ddba" (UID: "54ede1d0-752d-4967-a158-d41b7d56ddba"). InnerVolumeSpecName "kube-api-access-x5d7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:47 crc kubenswrapper[4696]: W0321 08:46:47.773984 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbac528ca_04da_4a92_9b88_b9a3d226cb94.slice/crio-7b7ca286e8b1e552a19d74a1bb6df332a31fde44386ece2befb092da58116b4c WatchSource:0}: Error finding container 7b7ca286e8b1e552a19d74a1bb6df332a31fde44386ece2befb092da58116b4c: Status 404 returned error can't find the container with id 7b7ca286e8b1e552a19d74a1bb6df332a31fde44386ece2befb092da58116b4c Mar 21 08:46:47 crc kubenswrapper[4696]: W0321 08:46:47.775390 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a6217ca_7143_41fd_ae0d_f96bde35029c.slice/crio-77a402cb517713352ad4a0ec0b8f6dc314e6b299dadfe465e151608c57e7223a WatchSource:0}: Error finding container 77a402cb517713352ad4a0ec0b8f6dc314e6b299dadfe465e151608c57e7223a: Status 404 returned error can't find the container with id 77a402cb517713352ad4a0ec0b8f6dc314e6b299dadfe465e151608c57e7223a Mar 21 08:46:47 crc kubenswrapper[4696]: W0321 08:46:47.780757 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod80468ce8_5377_4423_9e04_b58d1d0910d7.slice/crio-2f8cb46e168dd3c1b6fff9ffc8d3fc179356143a33914352c6e3196c07704a3d WatchSource:0}: Error finding container 2f8cb46e168dd3c1b6fff9ffc8d3fc179356143a33914352c6e3196c07704a3d: Status 404 returned error can't find the container with id 2f8cb46e168dd3c1b6fff9ffc8d3fc179356143a33914352c6e3196c07704a3d Mar 21 08:46:47 crc kubenswrapper[4696]: I0321 08:46:47.786390 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Mar 21 08:46:47 crc kubenswrapper[4696]: I0321 08:46:47.786919 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Mar 21 08:46:47 crc kubenswrapper[4696]: W0321 08:46:47.788117 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92718cf1_7a8f_42cd_aadc_df800f2499d4.slice/crio-ade4e3d38b01521263bfd605cfcb97d9bb753d4d0764db1b8aac05dcdd99be0c WatchSource:0}: Error finding container ade4e3d38b01521263bfd605cfcb97d9bb753d4d0764db1b8aac05dcdd99be0c: Status 404 returned error can't find the container with id ade4e3d38b01521263bfd605cfcb97d9bb753d4d0764db1b8aac05dcdd99be0c Mar 21 08:46:47 crc kubenswrapper[4696]: I0321 08:46:47.796626 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5d7m\" (UniqueName: \"kubernetes.io/projected/54ede1d0-752d-4967-a158-d41b7d56ddba-kube-api-access-x5d7m\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:47 crc kubenswrapper[4696]: I0321 08:46:47.796650 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54ede1d0-752d-4967-a158-d41b7d56ddba-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:47 crc kubenswrapper[4696]: I0321 08:46:47.898045 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-2ef2-account-create-update-9pxpg" Mar 21 08:46:47 crc kubenswrapper[4696]: I0321 08:46:47.908357 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-mtbts" Mar 21 08:46:47 crc kubenswrapper[4696]: I0321 08:46:47.952504 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f359-account-create-update-2jqvh" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.001169 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vklxz\" (UniqueName: \"kubernetes.io/projected/17520989-db17-4ad3-8324-279698af93e3-kube-api-access-vklxz\") pod \"17520989-db17-4ad3-8324-279698af93e3\" (UID: \"17520989-db17-4ad3-8324-279698af93e3\") " Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.001242 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17520989-db17-4ad3-8324-279698af93e3-operator-scripts\") pod \"17520989-db17-4ad3-8324-279698af93e3\" (UID: \"17520989-db17-4ad3-8324-279698af93e3\") " Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.001987 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17520989-db17-4ad3-8324-279698af93e3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "17520989-db17-4ad3-8324-279698af93e3" (UID: "17520989-db17-4ad3-8324-279698af93e3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.002298 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17520989-db17-4ad3-8324-279698af93e3-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.008458 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17520989-db17-4ad3-8324-279698af93e3-kube-api-access-vklxz" (OuterVolumeSpecName: "kube-api-access-vklxz") pod "17520989-db17-4ad3-8324-279698af93e3" (UID: "17520989-db17-4ad3-8324-279698af93e3"). InnerVolumeSpecName "kube-api-access-vklxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.103830 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmmjg\" (UniqueName: \"kubernetes.io/projected/5be1b1b8-ee1e-47cd-afbe-e1a96cee697a-kube-api-access-cmmjg\") pod \"5be1b1b8-ee1e-47cd-afbe-e1a96cee697a\" (UID: \"5be1b1b8-ee1e-47cd-afbe-e1a96cee697a\") " Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.104381 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec226d0f-1ee8-481b-85a5-280fe2d17810-operator-scripts\") pod \"ec226d0f-1ee8-481b-85a5-280fe2d17810\" (UID: \"ec226d0f-1ee8-481b-85a5-280fe2d17810\") " Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.104628 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6d5n\" (UniqueName: \"kubernetes.io/projected/ec226d0f-1ee8-481b-85a5-280fe2d17810-kube-api-access-j6d5n\") pod \"ec226d0f-1ee8-481b-85a5-280fe2d17810\" (UID: \"ec226d0f-1ee8-481b-85a5-280fe2d17810\") " Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.104775 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5be1b1b8-ee1e-47cd-afbe-e1a96cee697a-operator-scripts\") pod \"5be1b1b8-ee1e-47cd-afbe-e1a96cee697a\" (UID: \"5be1b1b8-ee1e-47cd-afbe-e1a96cee697a\") " Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.104835 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec226d0f-1ee8-481b-85a5-280fe2d17810-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ec226d0f-1ee8-481b-85a5-280fe2d17810" (UID: "ec226d0f-1ee8-481b-85a5-280fe2d17810"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.105119 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5be1b1b8-ee1e-47cd-afbe-e1a96cee697a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5be1b1b8-ee1e-47cd-afbe-e1a96cee697a" (UID: "5be1b1b8-ee1e-47cd-afbe-e1a96cee697a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.105442 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vklxz\" (UniqueName: \"kubernetes.io/projected/17520989-db17-4ad3-8324-279698af93e3-kube-api-access-vklxz\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.105516 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec226d0f-1ee8-481b-85a5-280fe2d17810-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.105577 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5be1b1b8-ee1e-47cd-afbe-e1a96cee697a-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.109866 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec226d0f-1ee8-481b-85a5-280fe2d17810-kube-api-access-j6d5n" (OuterVolumeSpecName: "kube-api-access-j6d5n") pod "ec226d0f-1ee8-481b-85a5-280fe2d17810" (UID: "ec226d0f-1ee8-481b-85a5-280fe2d17810"). InnerVolumeSpecName "kube-api-access-j6d5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.116668 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5be1b1b8-ee1e-47cd-afbe-e1a96cee697a-kube-api-access-cmmjg" (OuterVolumeSpecName: "kube-api-access-cmmjg") pod "5be1b1b8-ee1e-47cd-afbe-e1a96cee697a" (UID: "5be1b1b8-ee1e-47cd-afbe-e1a96cee697a"). InnerVolumeSpecName "kube-api-access-cmmjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.200829 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-mtbts" event={"ID":"5be1b1b8-ee1e-47cd-afbe-e1a96cee697a","Type":"ContainerDied","Data":"205add55b36fb047113d7717db8ce5a124cf5adadf9e2fc023fcd7ff0c89c3fa"} Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.200881 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="205add55b36fb047113d7717db8ce5a124cf5adadf9e2fc023fcd7ff0c89c3fa" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.200940 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-mtbts" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.209376 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmmjg\" (UniqueName: \"kubernetes.io/projected/5be1b1b8-ee1e-47cd-afbe-e1a96cee697a-kube-api-access-cmmjg\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.209399 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6d5n\" (UniqueName: \"kubernetes.io/projected/ec226d0f-1ee8-481b-85a5-280fe2d17810-kube-api-access-j6d5n\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.210588 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-52e3-account-create-update-5pmnh" event={"ID":"bac528ca-04da-4a92-9b88-b9a3d226cb94","Type":"ContainerStarted","Data":"1c41dad6f5faecac40e3d9cc1e1ff2d96ab5786733f4983487418a99056fe5f5"} Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.210626 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-52e3-account-create-update-5pmnh" event={"ID":"bac528ca-04da-4a92-9b88-b9a3d226cb94","Type":"ContainerStarted","Data":"7b7ca286e8b1e552a19d74a1bb6df332a31fde44386ece2befb092da58116b4c"} Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.220936 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-tjmmf" event={"ID":"cf260a42-d042-4130-bc96-9e12dcb253c4","Type":"ContainerStarted","Data":"95500aa101eb2803089e3ba8b625535edbe027bab63443bf2bfd2c0e85e980df"} Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.230558 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-trn7c" event={"ID":"92718cf1-7a8f-42cd-aadc-df800f2499d4","Type":"ContainerStarted","Data":"ade4e3d38b01521263bfd605cfcb97d9bb753d4d0764db1b8aac05dcdd99be0c"} Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.231755 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f359-account-create-update-2jqvh" event={"ID":"ec226d0f-1ee8-481b-85a5-280fe2d17810","Type":"ContainerDied","Data":"70db639e7e34186303d1a788ca83dbcce42d34c2d2454c9312209ba35666a955"} Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.231781 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70db639e7e34186303d1a788ca83dbcce42d34c2d2454c9312209ba35666a955" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.231851 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f359-account-create-update-2jqvh" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.241154 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-52e3-account-create-update-5pmnh" podStartSLOduration=10.24112937 podStartE2EDuration="10.24112937s" podCreationTimestamp="2026-03-21 08:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:46:48.228626011 +0000 UTC m=+1142.349506724" watchObservedRunningTime="2026-03-21 08:46:48.24112937 +0000 UTC m=+1142.362010083" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.250223 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-2ef2-account-create-update-9pxpg" event={"ID":"17520989-db17-4ad3-8324-279698af93e3","Type":"ContainerDied","Data":"ca2886eb12cee6010f9f5dddb018051353dae74350b01dfee2b1ced498bdbbcd"} Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.250258 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca2886eb12cee6010f9f5dddb018051353dae74350b01dfee2b1ced498bdbbcd" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.250330 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-2ef2-account-create-update-9pxpg" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.258402 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-db-create-tjmmf" podStartSLOduration=10.258375781 podStartE2EDuration="10.258375781s" podCreationTimestamp="2026-03-21 08:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:46:48.247612411 +0000 UTC m=+1142.368493124" watchObservedRunningTime="2026-03-21 08:46:48.258375781 +0000 UTC m=+1142.379256494" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.265345 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4098-account-create-update-mn786" event={"ID":"1a6217ca-7143-41fd-ae0d-f96bde35029c","Type":"ContainerStarted","Data":"b7568e0277590b09dcd295ac32141b6adf5f79f58c74f5c1897d7e2d67b2a912"} Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.265385 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4098-account-create-update-mn786" event={"ID":"1a6217ca-7143-41fd-ae0d-f96bde35029c","Type":"ContainerStarted","Data":"77a402cb517713352ad4a0ec0b8f6dc314e6b299dadfe465e151608c57e7223a"} Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.274140 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rff8n" event={"ID":"dc7fc65b-70b4-46cd-84e5-16285d93f295","Type":"ContainerStarted","Data":"7aaf2ff35fada122fcd81803dbb0c9200193736cb8409fff9c307db1291cbdf4"} Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.282478 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-b2dkt" event={"ID":"54ede1d0-752d-4967-a158-d41b7d56ddba","Type":"ContainerDied","Data":"dc821e08b634b2f9bea2956ee7fa96ae605788a3abcfa6ee85c1d7ec3e1ac768"} Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.282525 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc821e08b634b2f9bea2956ee7fa96ae605788a3abcfa6ee85c1d7ec3e1ac768" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.282593 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-b2dkt" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.298077 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"c2c63b45-de34-43a6-bb8b-aed4b00922c8","Type":"ContainerStarted","Data":"4c1ceb9af45601052e342e6c9eda973a9335cf70e88c88394897d1e58826843c"} Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.302701 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-jc8bj" event={"ID":"80468ce8-5377-4423-9e04-b58d1d0910d7","Type":"ContainerStarted","Data":"f0b2b6e33fa875b269cbc9bceec63983ad8d84467cc4ec30c19f68f4d7c85243"} Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.302851 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-jc8bj" event={"ID":"80468ce8-5377-4423-9e04-b58d1d0910d7","Type":"ContainerStarted","Data":"2f8cb46e168dd3c1b6fff9ffc8d3fc179356143a33914352c6e3196c07704a3d"} Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.327186 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-rff8n" podStartSLOduration=10.327168782 podStartE2EDuration="10.327168782s" podCreationTimestamp="2026-03-21 08:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:46:48.326201184 +0000 UTC m=+1142.447081897" watchObservedRunningTime="2026-03-21 08:46:48.327168782 +0000 UTC m=+1142.448049495" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.334653 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-4098-account-create-update-mn786" podStartSLOduration=10.33463465 podStartE2EDuration="10.33463465s" podCreationTimestamp="2026-03-21 08:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:46:48.303104779 +0000 UTC m=+1142.423985592" watchObservedRunningTime="2026-03-21 08:46:48.33463465 +0000 UTC m=+1142.455515363" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.357155 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=20.357135627 podStartE2EDuration="20.357135627s" podCreationTimestamp="2026-03-21 08:46:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:46:48.347770336 +0000 UTC m=+1142.468651059" watchObservedRunningTime="2026-03-21 08:46:48.357135627 +0000 UTC m=+1142.478016340" Mar 21 08:46:48 crc kubenswrapper[4696]: I0321 08:46:48.368414 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-jc8bj" podStartSLOduration=10.368393552 podStartE2EDuration="10.368393552s" podCreationTimestamp="2026-03-21 08:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:46:48.366138788 +0000 UTC m=+1142.487019521" watchObservedRunningTime="2026-03-21 08:46:48.368393552 +0000 UTC m=+1142.489274265" Mar 21 08:46:49 crc kubenswrapper[4696]: I0321 08:46:49.283918 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:49 crc kubenswrapper[4696]: I0321 08:46:49.313291 4696 generic.go:334] "Generic (PLEG): container finished" podID="dc7fc65b-70b4-46cd-84e5-16285d93f295" containerID="7aaf2ff35fada122fcd81803dbb0c9200193736cb8409fff9c307db1291cbdf4" exitCode=0 Mar 21 08:46:49 crc kubenswrapper[4696]: I0321 08:46:49.313725 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rff8n" event={"ID":"dc7fc65b-70b4-46cd-84e5-16285d93f295","Type":"ContainerDied","Data":"7aaf2ff35fada122fcd81803dbb0c9200193736cb8409fff9c307db1291cbdf4"} Mar 21 08:46:49 crc kubenswrapper[4696]: I0321 08:46:49.317137 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1088ee04-73e8-4aba-aad4-1cd7d26e20aa","Type":"ContainerStarted","Data":"3cf4315eb54c3756e57bcc965cc73329ebaa33667037cccea58540a5d592beb3"} Mar 21 08:46:49 crc kubenswrapper[4696]: I0321 08:46:49.317181 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1088ee04-73e8-4aba-aad4-1cd7d26e20aa","Type":"ContainerStarted","Data":"ce36f07b11026bfc663b01be3e47c23e467536b7e63c83fdcc6694407df7e9e7"} Mar 21 08:46:49 crc kubenswrapper[4696]: I0321 08:46:49.317194 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1088ee04-73e8-4aba-aad4-1cd7d26e20aa","Type":"ContainerStarted","Data":"5bf45649f95265e5f0c6f09827f9a0c55fe0175741cab574e828eeeffbe8f525"} Mar 21 08:46:49 crc kubenswrapper[4696]: I0321 08:46:49.317206 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1088ee04-73e8-4aba-aad4-1cd7d26e20aa","Type":"ContainerStarted","Data":"5ead6bf0e4b20b3d91edf03939f430feafbecce403c3e3a650a7cf68d39ddc28"} Mar 21 08:46:49 crc kubenswrapper[4696]: I0321 08:46:49.318980 4696 generic.go:334] "Generic (PLEG): container finished" podID="80468ce8-5377-4423-9e04-b58d1d0910d7" containerID="f0b2b6e33fa875b269cbc9bceec63983ad8d84467cc4ec30c19f68f4d7c85243" exitCode=0 Mar 21 08:46:49 crc kubenswrapper[4696]: I0321 08:46:49.319040 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-jc8bj" event={"ID":"80468ce8-5377-4423-9e04-b58d1d0910d7","Type":"ContainerDied","Data":"f0b2b6e33fa875b269cbc9bceec63983ad8d84467cc4ec30c19f68f4d7c85243"} Mar 21 08:46:49 crc kubenswrapper[4696]: I0321 08:46:49.324073 4696 generic.go:334] "Generic (PLEG): container finished" podID="cf260a42-d042-4130-bc96-9e12dcb253c4" containerID="95500aa101eb2803089e3ba8b625535edbe027bab63443bf2bfd2c0e85e980df" exitCode=0 Mar 21 08:46:49 crc kubenswrapper[4696]: I0321 08:46:49.324174 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-tjmmf" event={"ID":"cf260a42-d042-4130-bc96-9e12dcb253c4","Type":"ContainerDied","Data":"95500aa101eb2803089e3ba8b625535edbe027bab63443bf2bfd2c0e85e980df"} Mar 21 08:46:49 crc kubenswrapper[4696]: I0321 08:46:49.326236 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jqpz7" event={"ID":"7359c71c-56ba-4c28-a126-bc95dea18dfd","Type":"ContainerStarted","Data":"39e6b99fa4de32a798f582ebba0755776e9fe76d1fef544f75ba6269ef3aa353"} Mar 21 08:46:49 crc kubenswrapper[4696]: I0321 08:46:49.327660 4696 generic.go:334] "Generic (PLEG): container finished" podID="bac528ca-04da-4a92-9b88-b9a3d226cb94" containerID="1c41dad6f5faecac40e3d9cc1e1ff2d96ab5786733f4983487418a99056fe5f5" exitCode=0 Mar 21 08:46:49 crc kubenswrapper[4696]: I0321 08:46:49.327706 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-52e3-account-create-update-5pmnh" event={"ID":"bac528ca-04da-4a92-9b88-b9a3d226cb94","Type":"ContainerDied","Data":"1c41dad6f5faecac40e3d9cc1e1ff2d96ab5786733f4983487418a99056fe5f5"} Mar 21 08:46:49 crc kubenswrapper[4696]: I0321 08:46:49.329298 4696 generic.go:334] "Generic (PLEG): container finished" podID="1a6217ca-7143-41fd-ae0d-f96bde35029c" containerID="b7568e0277590b09dcd295ac32141b6adf5f79f58c74f5c1897d7e2d67b2a912" exitCode=0 Mar 21 08:46:49 crc kubenswrapper[4696]: I0321 08:46:49.329332 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4098-account-create-update-mn786" event={"ID":"1a6217ca-7143-41fd-ae0d-f96bde35029c","Type":"ContainerDied","Data":"b7568e0277590b09dcd295ac32141b6adf5f79f58c74f5c1897d7e2d67b2a912"} Mar 21 08:46:49 crc kubenswrapper[4696]: I0321 08:46:49.377971 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-jqpz7" podStartSLOduration=3.014791266 podStartE2EDuration="20.377953494s" podCreationTimestamp="2026-03-21 08:46:29 +0000 UTC" firstStartedPulling="2026-03-21 08:46:30.631885045 +0000 UTC m=+1124.752765748" lastFinishedPulling="2026-03-21 08:46:47.995047263 +0000 UTC m=+1142.115927976" observedRunningTime="2026-03-21 08:46:49.375833195 +0000 UTC m=+1143.496713918" watchObservedRunningTime="2026-03-21 08:46:49.377953494 +0000 UTC m=+1143.498834227" Mar 21 08:46:50 crc kubenswrapper[4696]: I0321 08:46:50.658290 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-ingester-0" Mar 21 08:46:50 crc kubenswrapper[4696]: I0321 08:46:50.798171 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4098-account-create-update-mn786" Mar 21 08:46:50 crc kubenswrapper[4696]: I0321 08:46:50.881325 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mn44v\" (UniqueName: \"kubernetes.io/projected/1a6217ca-7143-41fd-ae0d-f96bde35029c-kube-api-access-mn44v\") pod \"1a6217ca-7143-41fd-ae0d-f96bde35029c\" (UID: \"1a6217ca-7143-41fd-ae0d-f96bde35029c\") " Mar 21 08:46:50 crc kubenswrapper[4696]: I0321 08:46:50.881444 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a6217ca-7143-41fd-ae0d-f96bde35029c-operator-scripts\") pod \"1a6217ca-7143-41fd-ae0d-f96bde35029c\" (UID: \"1a6217ca-7143-41fd-ae0d-f96bde35029c\") " Mar 21 08:46:50 crc kubenswrapper[4696]: I0321 08:46:50.882588 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a6217ca-7143-41fd-ae0d-f96bde35029c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1a6217ca-7143-41fd-ae0d-f96bde35029c" (UID: "1a6217ca-7143-41fd-ae0d-f96bde35029c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:50 crc kubenswrapper[4696]: I0321 08:46:50.895293 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a6217ca-7143-41fd-ae0d-f96bde35029c-kube-api-access-mn44v" (OuterVolumeSpecName: "kube-api-access-mn44v") pod "1a6217ca-7143-41fd-ae0d-f96bde35029c" (UID: "1a6217ca-7143-41fd-ae0d-f96bde35029c"). InnerVolumeSpecName "kube-api-access-mn44v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:50 crc kubenswrapper[4696]: I0321 08:46:50.946921 4696 scope.go:117] "RemoveContainer" containerID="94325b9c899893b0d183be3caa5acc3b05b5cdcc90d52a9b47f06593563370b2" Mar 21 08:46:50 crc kubenswrapper[4696]: I0321 08:46:50.983796 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a6217ca-7143-41fd-ae0d-f96bde35029c-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:50 crc kubenswrapper[4696]: I0321 08:46:50.983840 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mn44v\" (UniqueName: \"kubernetes.io/projected/1a6217ca-7143-41fd-ae0d-f96bde35029c-kube-api-access-mn44v\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.144368 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jc8bj" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.158875 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-tjmmf" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.160798 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rff8n" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.196458 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-52e3-account-create-update-5pmnh" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.298649 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc7fc65b-70b4-46cd-84e5-16285d93f295-operator-scripts\") pod \"dc7fc65b-70b4-46cd-84e5-16285d93f295\" (UID: \"dc7fc65b-70b4-46cd-84e5-16285d93f295\") " Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.298731 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf260a42-d042-4130-bc96-9e12dcb253c4-operator-scripts\") pod \"cf260a42-d042-4130-bc96-9e12dcb253c4\" (UID: \"cf260a42-d042-4130-bc96-9e12dcb253c4\") " Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.298765 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80468ce8-5377-4423-9e04-b58d1d0910d7-operator-scripts\") pod \"80468ce8-5377-4423-9e04-b58d1d0910d7\" (UID: \"80468ce8-5377-4423-9e04-b58d1d0910d7\") " Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.298811 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bac528ca-04da-4a92-9b88-b9a3d226cb94-operator-scripts\") pod \"bac528ca-04da-4a92-9b88-b9a3d226cb94\" (UID: \"bac528ca-04da-4a92-9b88-b9a3d226cb94\") " Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.298896 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kppb\" (UniqueName: \"kubernetes.io/projected/bac528ca-04da-4a92-9b88-b9a3d226cb94-kube-api-access-9kppb\") pod \"bac528ca-04da-4a92-9b88-b9a3d226cb94\" (UID: \"bac528ca-04da-4a92-9b88-b9a3d226cb94\") " Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.298959 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wb455\" (UniqueName: \"kubernetes.io/projected/80468ce8-5377-4423-9e04-b58d1d0910d7-kube-api-access-wb455\") pod \"80468ce8-5377-4423-9e04-b58d1d0910d7\" (UID: \"80468ce8-5377-4423-9e04-b58d1d0910d7\") " Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.298986 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnfxq\" (UniqueName: \"kubernetes.io/projected/dc7fc65b-70b4-46cd-84e5-16285d93f295-kube-api-access-mnfxq\") pod \"dc7fc65b-70b4-46cd-84e5-16285d93f295\" (UID: \"dc7fc65b-70b4-46cd-84e5-16285d93f295\") " Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.299022 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tr8wq\" (UniqueName: \"kubernetes.io/projected/cf260a42-d042-4130-bc96-9e12dcb253c4-kube-api-access-tr8wq\") pod \"cf260a42-d042-4130-bc96-9e12dcb253c4\" (UID: \"cf260a42-d042-4130-bc96-9e12dcb253c4\") " Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.300900 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf260a42-d042-4130-bc96-9e12dcb253c4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cf260a42-d042-4130-bc96-9e12dcb253c4" (UID: "cf260a42-d042-4130-bc96-9e12dcb253c4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.300924 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc7fc65b-70b4-46cd-84e5-16285d93f295-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dc7fc65b-70b4-46cd-84e5-16285d93f295" (UID: "dc7fc65b-70b4-46cd-84e5-16285d93f295"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.301273 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80468ce8-5377-4423-9e04-b58d1d0910d7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "80468ce8-5377-4423-9e04-b58d1d0910d7" (UID: "80468ce8-5377-4423-9e04-b58d1d0910d7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.301775 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bac528ca-04da-4a92-9b88-b9a3d226cb94-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bac528ca-04da-4a92-9b88-b9a3d226cb94" (UID: "bac528ca-04da-4a92-9b88-b9a3d226cb94"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.304667 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bac528ca-04da-4a92-9b88-b9a3d226cb94-kube-api-access-9kppb" (OuterVolumeSpecName: "kube-api-access-9kppb") pod "bac528ca-04da-4a92-9b88-b9a3d226cb94" (UID: "bac528ca-04da-4a92-9b88-b9a3d226cb94"). InnerVolumeSpecName "kube-api-access-9kppb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.304706 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf260a42-d042-4130-bc96-9e12dcb253c4-kube-api-access-tr8wq" (OuterVolumeSpecName: "kube-api-access-tr8wq") pod "cf260a42-d042-4130-bc96-9e12dcb253c4" (UID: "cf260a42-d042-4130-bc96-9e12dcb253c4"). InnerVolumeSpecName "kube-api-access-tr8wq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.305556 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc7fc65b-70b4-46cd-84e5-16285d93f295-kube-api-access-mnfxq" (OuterVolumeSpecName: "kube-api-access-mnfxq") pod "dc7fc65b-70b4-46cd-84e5-16285d93f295" (UID: "dc7fc65b-70b4-46cd-84e5-16285d93f295"). InnerVolumeSpecName "kube-api-access-mnfxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.306023 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80468ce8-5377-4423-9e04-b58d1d0910d7-kube-api-access-wb455" (OuterVolumeSpecName: "kube-api-access-wb455") pod "80468ce8-5377-4423-9e04-b58d1d0910d7" (UID: "80468ce8-5377-4423-9e04-b58d1d0910d7"). InnerVolumeSpecName "kube-api-access-wb455". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.344788 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jc8bj" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.344920 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-jc8bj" event={"ID":"80468ce8-5377-4423-9e04-b58d1d0910d7","Type":"ContainerDied","Data":"2f8cb46e168dd3c1b6fff9ffc8d3fc179356143a33914352c6e3196c07704a3d"} Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.344967 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f8cb46e168dd3c1b6fff9ffc8d3fc179356143a33914352c6e3196c07704a3d" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.346132 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4098-account-create-update-mn786" event={"ID":"1a6217ca-7143-41fd-ae0d-f96bde35029c","Type":"ContainerDied","Data":"77a402cb517713352ad4a0ec0b8f6dc314e6b299dadfe465e151608c57e7223a"} Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.346242 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4098-account-create-update-mn786" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.346263 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77a402cb517713352ad4a0ec0b8f6dc314e6b299dadfe465e151608c57e7223a" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.347146 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rff8n" event={"ID":"dc7fc65b-70b4-46cd-84e5-16285d93f295","Type":"ContainerDied","Data":"c63dee5f82e8064bc80b8774e641cfdff9efe69ef0ac4f5b4ff5e03b266fe549"} Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.347164 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c63dee5f82e8064bc80b8774e641cfdff9efe69ef0ac4f5b4ff5e03b266fe549" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.347207 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rff8n" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.357528 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-tjmmf" event={"ID":"cf260a42-d042-4130-bc96-9e12dcb253c4","Type":"ContainerDied","Data":"da3e06e5c2261fd2a3d9b4990569f2bf8c3f9ac0bca2060fee118dd083e2b3d3"} Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.357565 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da3e06e5c2261fd2a3d9b4990569f2bf8c3f9ac0bca2060fee118dd083e2b3d3" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.357573 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-tjmmf" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.361883 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-52e3-account-create-update-5pmnh" event={"ID":"bac528ca-04da-4a92-9b88-b9a3d226cb94","Type":"ContainerDied","Data":"7b7ca286e8b1e552a19d74a1bb6df332a31fde44386ece2befb092da58116b4c"} Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.361953 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-52e3-account-create-update-5pmnh" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.361965 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b7ca286e8b1e552a19d74a1bb6df332a31fde44386ece2befb092da58116b4c" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.400974 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tr8wq\" (UniqueName: \"kubernetes.io/projected/cf260a42-d042-4130-bc96-9e12dcb253c4-kube-api-access-tr8wq\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.401242 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc7fc65b-70b4-46cd-84e5-16285d93f295-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.401362 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf260a42-d042-4130-bc96-9e12dcb253c4-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.401418 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80468ce8-5377-4423-9e04-b58d1d0910d7-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.401478 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bac528ca-04da-4a92-9b88-b9a3d226cb94-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.401568 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kppb\" (UniqueName: \"kubernetes.io/projected/bac528ca-04da-4a92-9b88-b9a3d226cb94-kube-api-access-9kppb\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.401697 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wb455\" (UniqueName: \"kubernetes.io/projected/80468ce8-5377-4423-9e04-b58d1d0910d7-kube-api-access-wb455\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:51 crc kubenswrapper[4696]: I0321 08:46:51.401876 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnfxq\" (UniqueName: \"kubernetes.io/projected/dc7fc65b-70b4-46cd-84e5-16285d93f295-kube-api-access-mnfxq\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:52 crc kubenswrapper[4696]: I0321 08:46:52.562691 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-b2dkt"] Mar 21 08:46:52 crc kubenswrapper[4696]: I0321 08:46:52.570770 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-b2dkt"] Mar 21 08:46:53 crc kubenswrapper[4696]: I0321 08:46:53.384546 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1088ee04-73e8-4aba-aad4-1cd7d26e20aa","Type":"ContainerStarted","Data":"0a9a8ac6ec7da60cc0c65cbb4a1f0f435053c2a5f8d5bc30ff0613cd5b2b2d84"} Mar 21 08:46:53 crc kubenswrapper[4696]: I0321 08:46:53.384889 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1088ee04-73e8-4aba-aad4-1cd7d26e20aa","Type":"ContainerStarted","Data":"2a9a2fd851fc3362d5224a3e84e85057d5cd4212744acfe530be1baa57e4b5bf"} Mar 21 08:46:53 crc kubenswrapper[4696]: I0321 08:46:53.384904 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1088ee04-73e8-4aba-aad4-1cd7d26e20aa","Type":"ContainerStarted","Data":"acfd7f91ec98775ef5559aef601b03dafa6d42f90949ca674c466da04f0ba721"} Mar 21 08:46:53 crc kubenswrapper[4696]: I0321 08:46:53.384917 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1088ee04-73e8-4aba-aad4-1cd7d26e20aa","Type":"ContainerStarted","Data":"16109740180d75eb16bcfc4c968b060fe19349a11c011138b4fb56c59e3857b2"} Mar 21 08:46:54 crc kubenswrapper[4696]: I0321 08:46:54.547104 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54ede1d0-752d-4967-a158-d41b7d56ddba" path="/var/lib/kubelet/pods/54ede1d0-752d-4967-a158-d41b7d56ddba/volumes" Mar 21 08:46:56 crc kubenswrapper[4696]: I0321 08:46:56.413951 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-trn7c" event={"ID":"92718cf1-7a8f-42cd-aadc-df800f2499d4","Type":"ContainerStarted","Data":"23bd157c07ed6f47aae360143f80c9569b30775dd932dfdee1c2a28699730de9"} Mar 21 08:46:56 crc kubenswrapper[4696]: I0321 08:46:56.430459 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1088ee04-73e8-4aba-aad4-1cd7d26e20aa","Type":"ContainerStarted","Data":"4d3f1e6420516ba040ecf727a7b57eb123e535aa4490c44ca73f37bfb1c8f013"} Mar 21 08:46:56 crc kubenswrapper[4696]: I0321 08:46:56.430509 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1088ee04-73e8-4aba-aad4-1cd7d26e20aa","Type":"ContainerStarted","Data":"221339c17171f25b2ef9909125144d509d2cf7c5d3a8c8f17f95e94e7f9685ce"} Mar 21 08:46:56 crc kubenswrapper[4696]: I0321 08:46:56.431315 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-trn7c" podStartSLOduration=10.323457817 podStartE2EDuration="18.43127367s" podCreationTimestamp="2026-03-21 08:46:38 +0000 UTC" firstStartedPulling="2026-03-21 08:46:47.85770688 +0000 UTC m=+1141.978587593" lastFinishedPulling="2026-03-21 08:46:55.965522733 +0000 UTC m=+1150.086403446" observedRunningTime="2026-03-21 08:46:56.430337714 +0000 UTC m=+1150.551218447" watchObservedRunningTime="2026-03-21 08:46:56.43127367 +0000 UTC m=+1150.552154373" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.445046 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1088ee04-73e8-4aba-aad4-1cd7d26e20aa","Type":"ContainerStarted","Data":"0bb388e8aecdad633a1d79fd611711023b4deb436be43255ee7c25cc5c6451c9"} Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.447491 4696 generic.go:334] "Generic (PLEG): container finished" podID="7359c71c-56ba-4c28-a126-bc95dea18dfd" containerID="39e6b99fa4de32a798f582ebba0755776e9fe76d1fef544f75ba6269ef3aa353" exitCode=0 Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.447554 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jqpz7" event={"ID":"7359c71c-56ba-4c28-a126-bc95dea18dfd","Type":"ContainerDied","Data":"39e6b99fa4de32a798f582ebba0755776e9fe76d1fef544f75ba6269ef3aa353"} Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.487114 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=39.457139545 podStartE2EDuration="56.487099104s" podCreationTimestamp="2026-03-21 08:46:01 +0000 UTC" firstStartedPulling="2026-03-21 08:46:35.158737009 +0000 UTC m=+1129.279617722" lastFinishedPulling="2026-03-21 08:46:52.188696568 +0000 UTC m=+1146.309577281" observedRunningTime="2026-03-21 08:46:57.48052523 +0000 UTC m=+1151.601405963" watchObservedRunningTime="2026-03-21 08:46:57.487099104 +0000 UTC m=+1151.607979817" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.609425 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-d2jgl"] Mar 21 08:46:57 crc kubenswrapper[4696]: E0321 08:46:57.610047 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf260a42-d042-4130-bc96-9e12dcb253c4" containerName="mariadb-database-create" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.610076 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf260a42-d042-4130-bc96-9e12dcb253c4" containerName="mariadb-database-create" Mar 21 08:46:57 crc kubenswrapper[4696]: E0321 08:46:57.610092 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a6217ca-7143-41fd-ae0d-f96bde35029c" containerName="mariadb-account-create-update" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.610098 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a6217ca-7143-41fd-ae0d-f96bde35029c" containerName="mariadb-account-create-update" Mar 21 08:46:57 crc kubenswrapper[4696]: E0321 08:46:57.610114 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80468ce8-5377-4423-9e04-b58d1d0910d7" containerName="mariadb-database-create" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.610120 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="80468ce8-5377-4423-9e04-b58d1d0910d7" containerName="mariadb-database-create" Mar 21 08:46:57 crc kubenswrapper[4696]: E0321 08:46:57.610144 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bac528ca-04da-4a92-9b88-b9a3d226cb94" containerName="mariadb-account-create-update" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.610150 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="bac528ca-04da-4a92-9b88-b9a3d226cb94" containerName="mariadb-account-create-update" Mar 21 08:46:57 crc kubenswrapper[4696]: E0321 08:46:57.610168 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec226d0f-1ee8-481b-85a5-280fe2d17810" containerName="mariadb-account-create-update" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.610174 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec226d0f-1ee8-481b-85a5-280fe2d17810" containerName="mariadb-account-create-update" Mar 21 08:46:57 crc kubenswrapper[4696]: E0321 08:46:57.610190 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5be1b1b8-ee1e-47cd-afbe-e1a96cee697a" containerName="mariadb-database-create" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.610195 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="5be1b1b8-ee1e-47cd-afbe-e1a96cee697a" containerName="mariadb-database-create" Mar 21 08:46:57 crc kubenswrapper[4696]: E0321 08:46:57.610208 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54ede1d0-752d-4967-a158-d41b7d56ddba" containerName="mariadb-account-create-update" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.610213 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="54ede1d0-752d-4967-a158-d41b7d56ddba" containerName="mariadb-account-create-update" Mar 21 08:46:57 crc kubenswrapper[4696]: E0321 08:46:57.610240 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17520989-db17-4ad3-8324-279698af93e3" containerName="mariadb-account-create-update" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.610246 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="17520989-db17-4ad3-8324-279698af93e3" containerName="mariadb-account-create-update" Mar 21 08:46:57 crc kubenswrapper[4696]: E0321 08:46:57.610259 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc7fc65b-70b4-46cd-84e5-16285d93f295" containerName="mariadb-database-create" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.610265 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc7fc65b-70b4-46cd-84e5-16285d93f295" containerName="mariadb-database-create" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.610539 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec226d0f-1ee8-481b-85a5-280fe2d17810" containerName="mariadb-account-create-update" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.610568 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc7fc65b-70b4-46cd-84e5-16285d93f295" containerName="mariadb-database-create" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.610588 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf260a42-d042-4130-bc96-9e12dcb253c4" containerName="mariadb-database-create" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.610611 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="bac528ca-04da-4a92-9b88-b9a3d226cb94" containerName="mariadb-account-create-update" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.610620 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a6217ca-7143-41fd-ae0d-f96bde35029c" containerName="mariadb-account-create-update" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.610631 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="54ede1d0-752d-4967-a158-d41b7d56ddba" containerName="mariadb-account-create-update" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.610682 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="17520989-db17-4ad3-8324-279698af93e3" containerName="mariadb-account-create-update" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.610701 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="80468ce8-5377-4423-9e04-b58d1d0910d7" containerName="mariadb-database-create" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.610709 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="5be1b1b8-ee1e-47cd-afbe-e1a96cee697a" containerName="mariadb-database-create" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.611546 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-d2jgl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.626953 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-d2jgl"] Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.643462 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.719122 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78946b93-1c47-42b0-98b4-7a9454ea8dbc-operator-scripts\") pod \"root-account-create-update-d2jgl\" (UID: \"78946b93-1c47-42b0-98b4-7a9454ea8dbc\") " pod="openstack/root-account-create-update-d2jgl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.719195 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl7hr\" (UniqueName: \"kubernetes.io/projected/78946b93-1c47-42b0-98b4-7a9454ea8dbc-kube-api-access-bl7hr\") pod \"root-account-create-update-d2jgl\" (UID: \"78946b93-1c47-42b0-98b4-7a9454ea8dbc\") " pod="openstack/root-account-create-update-d2jgl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.749590 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-v6pnl"] Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.751188 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.752742 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.794444 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-v6pnl"] Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.820701 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-v6pnl\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.820853 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-config\") pod \"dnsmasq-dns-764c5664d7-v6pnl\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.820876 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-v6pnl\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.820951 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-v6pnl\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.821012 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78946b93-1c47-42b0-98b4-7a9454ea8dbc-operator-scripts\") pod \"root-account-create-update-d2jgl\" (UID: \"78946b93-1c47-42b0-98b4-7a9454ea8dbc\") " pod="openstack/root-account-create-update-d2jgl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.821072 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-dns-svc\") pod \"dnsmasq-dns-764c5664d7-v6pnl\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.821093 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzh27\" (UniqueName: \"kubernetes.io/projected/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-kube-api-access-dzh27\") pod \"dnsmasq-dns-764c5664d7-v6pnl\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.821774 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78946b93-1c47-42b0-98b4-7a9454ea8dbc-operator-scripts\") pod \"root-account-create-update-d2jgl\" (UID: \"78946b93-1c47-42b0-98b4-7a9454ea8dbc\") " pod="openstack/root-account-create-update-d2jgl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.821892 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl7hr\" (UniqueName: \"kubernetes.io/projected/78946b93-1c47-42b0-98b4-7a9454ea8dbc-kube-api-access-bl7hr\") pod \"root-account-create-update-d2jgl\" (UID: \"78946b93-1c47-42b0-98b4-7a9454ea8dbc\") " pod="openstack/root-account-create-update-d2jgl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.838722 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl7hr\" (UniqueName: \"kubernetes.io/projected/78946b93-1c47-42b0-98b4-7a9454ea8dbc-kube-api-access-bl7hr\") pod \"root-account-create-update-d2jgl\" (UID: \"78946b93-1c47-42b0-98b4-7a9454ea8dbc\") " pod="openstack/root-account-create-update-d2jgl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.923990 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzh27\" (UniqueName: \"kubernetes.io/projected/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-kube-api-access-dzh27\") pod \"dnsmasq-dns-764c5664d7-v6pnl\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.924459 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-v6pnl\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.924534 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-config\") pod \"dnsmasq-dns-764c5664d7-v6pnl\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.924560 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-v6pnl\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.924584 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-v6pnl\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.924638 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-dns-svc\") pod \"dnsmasq-dns-764c5664d7-v6pnl\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.925356 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-v6pnl\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.925438 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-config\") pod \"dnsmasq-dns-764c5664d7-v6pnl\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.925642 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-dns-svc\") pod \"dnsmasq-dns-764c5664d7-v6pnl\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.925772 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-v6pnl\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.925787 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-v6pnl\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.945893 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzh27\" (UniqueName: \"kubernetes.io/projected/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-kube-api-access-dzh27\") pod \"dnsmasq-dns-764c5664d7-v6pnl\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:46:57 crc kubenswrapper[4696]: I0321 08:46:57.976510 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-d2jgl" Mar 21 08:46:58 crc kubenswrapper[4696]: I0321 08:46:58.068814 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:46:58 crc kubenswrapper[4696]: W0321 08:46:58.249097 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78946b93_1c47_42b0_98b4_7a9454ea8dbc.slice/crio-af1e7011478ddfcb87c51458567f070d82bab729a02641c8c2dd986c654dc34c WatchSource:0}: Error finding container af1e7011478ddfcb87c51458567f070d82bab729a02641c8c2dd986c654dc34c: Status 404 returned error can't find the container with id af1e7011478ddfcb87c51458567f070d82bab729a02641c8c2dd986c654dc34c Mar 21 08:46:58 crc kubenswrapper[4696]: I0321 08:46:58.251128 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-d2jgl"] Mar 21 08:46:58 crc kubenswrapper[4696]: I0321 08:46:58.472558 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-d2jgl" event={"ID":"78946b93-1c47-42b0-98b4-7a9454ea8dbc","Type":"ContainerStarted","Data":"c43fd1da93fdfad7e2a19752084be7a45719936890db99fe36fff7b8af9b358e"} Mar 21 08:46:58 crc kubenswrapper[4696]: I0321 08:46:58.472610 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-d2jgl" event={"ID":"78946b93-1c47-42b0-98b4-7a9454ea8dbc","Type":"ContainerStarted","Data":"af1e7011478ddfcb87c51458567f070d82bab729a02641c8c2dd986c654dc34c"} Mar 21 08:46:58 crc kubenswrapper[4696]: I0321 08:46:58.489971 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-d2jgl" podStartSLOduration=1.489955669 podStartE2EDuration="1.489955669s" podCreationTimestamp="2026-03-21 08:46:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:46:58.486880633 +0000 UTC m=+1152.607761346" watchObservedRunningTime="2026-03-21 08:46:58.489955669 +0000 UTC m=+1152.610836382" Mar 21 08:46:58 crc kubenswrapper[4696]: I0321 08:46:58.591548 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-v6pnl"] Mar 21 08:46:58 crc kubenswrapper[4696]: W0321 08:46:58.593618 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ce3479f_1bdc_4779_9860_7fe9a59d6e94.slice/crio-e8c610b780fc56bec96452b55b4cad0181baff6aa276397f696edc73fc763d78 WatchSource:0}: Error finding container e8c610b780fc56bec96452b55b4cad0181baff6aa276397f696edc73fc763d78: Status 404 returned error can't find the container with id e8c610b780fc56bec96452b55b4cad0181baff6aa276397f696edc73fc763d78 Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.025035 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jqpz7" Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.068627 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7359c71c-56ba-4c28-a126-bc95dea18dfd-config-data\") pod \"7359c71c-56ba-4c28-a126-bc95dea18dfd\" (UID: \"7359c71c-56ba-4c28-a126-bc95dea18dfd\") " Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.068695 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7359c71c-56ba-4c28-a126-bc95dea18dfd-combined-ca-bundle\") pod \"7359c71c-56ba-4c28-a126-bc95dea18dfd\" (UID: \"7359c71c-56ba-4c28-a126-bc95dea18dfd\") " Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.068923 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7359c71c-56ba-4c28-a126-bc95dea18dfd-db-sync-config-data\") pod \"7359c71c-56ba-4c28-a126-bc95dea18dfd\" (UID: \"7359c71c-56ba-4c28-a126-bc95dea18dfd\") " Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.068962 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmb9h\" (UniqueName: \"kubernetes.io/projected/7359c71c-56ba-4c28-a126-bc95dea18dfd-kube-api-access-kmb9h\") pod \"7359c71c-56ba-4c28-a126-bc95dea18dfd\" (UID: \"7359c71c-56ba-4c28-a126-bc95dea18dfd\") " Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.075279 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7359c71c-56ba-4c28-a126-bc95dea18dfd-kube-api-access-kmb9h" (OuterVolumeSpecName: "kube-api-access-kmb9h") pod "7359c71c-56ba-4c28-a126-bc95dea18dfd" (UID: "7359c71c-56ba-4c28-a126-bc95dea18dfd"). InnerVolumeSpecName "kube-api-access-kmb9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.088746 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7359c71c-56ba-4c28-a126-bc95dea18dfd-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "7359c71c-56ba-4c28-a126-bc95dea18dfd" (UID: "7359c71c-56ba-4c28-a126-bc95dea18dfd"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.102935 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7359c71c-56ba-4c28-a126-bc95dea18dfd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7359c71c-56ba-4c28-a126-bc95dea18dfd" (UID: "7359c71c-56ba-4c28-a126-bc95dea18dfd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.130843 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7359c71c-56ba-4c28-a126-bc95dea18dfd-config-data" (OuterVolumeSpecName: "config-data") pod "7359c71c-56ba-4c28-a126-bc95dea18dfd" (UID: "7359c71c-56ba-4c28-a126-bc95dea18dfd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.170895 4696 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7359c71c-56ba-4c28-a126-bc95dea18dfd-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.170937 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmb9h\" (UniqueName: \"kubernetes.io/projected/7359c71c-56ba-4c28-a126-bc95dea18dfd-kube-api-access-kmb9h\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.170953 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7359c71c-56ba-4c28-a126-bc95dea18dfd-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.170965 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7359c71c-56ba-4c28-a126-bc95dea18dfd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.284078 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.295621 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.483682 4696 generic.go:334] "Generic (PLEG): container finished" podID="5ce3479f-1bdc-4779-9860-7fe9a59d6e94" containerID="f8977d91c22cb7cd0bc89772b38b637d0532cf700af1c078f99e5082fae85989" exitCode=0 Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.483754 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" event={"ID":"5ce3479f-1bdc-4779-9860-7fe9a59d6e94","Type":"ContainerDied","Data":"f8977d91c22cb7cd0bc89772b38b637d0532cf700af1c078f99e5082fae85989"} Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.483785 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" event={"ID":"5ce3479f-1bdc-4779-9860-7fe9a59d6e94","Type":"ContainerStarted","Data":"e8c610b780fc56bec96452b55b4cad0181baff6aa276397f696edc73fc763d78"} Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.485845 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-trn7c" event={"ID":"92718cf1-7a8f-42cd-aadc-df800f2499d4","Type":"ContainerDied","Data":"23bd157c07ed6f47aae360143f80c9569b30775dd932dfdee1c2a28699730de9"} Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.485811 4696 generic.go:334] "Generic (PLEG): container finished" podID="92718cf1-7a8f-42cd-aadc-df800f2499d4" containerID="23bd157c07ed6f47aae360143f80c9569b30775dd932dfdee1c2a28699730de9" exitCode=0 Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.487457 4696 generic.go:334] "Generic (PLEG): container finished" podID="78946b93-1c47-42b0-98b4-7a9454ea8dbc" containerID="c43fd1da93fdfad7e2a19752084be7a45719936890db99fe36fff7b8af9b358e" exitCode=0 Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.487500 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-d2jgl" event={"ID":"78946b93-1c47-42b0-98b4-7a9454ea8dbc","Type":"ContainerDied","Data":"c43fd1da93fdfad7e2a19752084be7a45719936890db99fe36fff7b8af9b358e"} Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.489046 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jqpz7" Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.489062 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jqpz7" event={"ID":"7359c71c-56ba-4c28-a126-bc95dea18dfd","Type":"ContainerDied","Data":"33f6e2facaef341cf101f0a2eb4651da0a268cdb95c8f7d9e9238691912feb19"} Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.489122 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33f6e2facaef341cf101f0a2eb4651da0a268cdb95c8f7d9e9238691912feb19" Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.520474 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Mar 21 08:46:59 crc kubenswrapper[4696]: I0321 08:46:59.939172 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-v6pnl"] Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.109550 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-bnf6p"] Mar 21 08:47:00 crc kubenswrapper[4696]: E0321 08:47:00.109939 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7359c71c-56ba-4c28-a126-bc95dea18dfd" containerName="glance-db-sync" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.109955 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7359c71c-56ba-4c28-a126-bc95dea18dfd" containerName="glance-db-sync" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.110132 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="7359c71c-56ba-4c28-a126-bc95dea18dfd" containerName="glance-db-sync" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.121679 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.134730 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-bnf6p"] Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.215089 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-bnf6p\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.215155 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-bnf6p\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.215213 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-bnf6p\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.215245 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-config\") pod \"dnsmasq-dns-74f6bcbc87-bnf6p\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.215266 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjd98\" (UniqueName: \"kubernetes.io/projected/7ec50e79-4227-4770-b7bc-7dd5f4178fad-kube-api-access-fjd98\") pod \"dnsmasq-dns-74f6bcbc87-bnf6p\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.215405 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-bnf6p\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.316909 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-bnf6p\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.316964 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-bnf6p\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.316985 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-bnf6p\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.317022 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-bnf6p\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.317048 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-config\") pod \"dnsmasq-dns-74f6bcbc87-bnf6p\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.317066 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjd98\" (UniqueName: \"kubernetes.io/projected/7ec50e79-4227-4770-b7bc-7dd5f4178fad-kube-api-access-fjd98\") pod \"dnsmasq-dns-74f6bcbc87-bnf6p\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.318114 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-bnf6p\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.318155 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-bnf6p\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.318239 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-bnf6p\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.318249 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-config\") pod \"dnsmasq-dns-74f6bcbc87-bnf6p\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.318527 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-bnf6p\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.342185 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.342242 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.345123 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjd98\" (UniqueName: \"kubernetes.io/projected/7ec50e79-4227-4770-b7bc-7dd5f4178fad-kube-api-access-fjd98\") pod \"dnsmasq-dns-74f6bcbc87-bnf6p\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.449031 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.503617 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" event={"ID":"5ce3479f-1bdc-4779-9860-7fe9a59d6e94","Type":"ContainerStarted","Data":"a95ef0ed856984fc1e9f6ad813bed8db93b34bf18988bcd1f080a5078c1422c2"} Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.531039 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" podStartSLOduration=3.531021576 podStartE2EDuration="3.531021576s" podCreationTimestamp="2026-03-21 08:46:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:47:00.525870512 +0000 UTC m=+1154.646751225" watchObservedRunningTime="2026-03-21 08:47:00.531021576 +0000 UTC m=+1154.651902289" Mar 21 08:47:00 crc kubenswrapper[4696]: I0321 08:47:00.906839 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-bnf6p"] Mar 21 08:47:00 crc kubenswrapper[4696]: W0321 08:47:00.966658 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ec50e79_4227_4770_b7bc_7dd5f4178fad.slice/crio-1ed07f806773bf783d387a27d94bd1dea29d39cc051966a924066d323a143e8d WatchSource:0}: Error finding container 1ed07f806773bf783d387a27d94bd1dea29d39cc051966a924066d323a143e8d: Status 404 returned error can't find the container with id 1ed07f806773bf783d387a27d94bd1dea29d39cc051966a924066d323a143e8d Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.141688 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-trn7c" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.150755 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-d2jgl" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.240154 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92718cf1-7a8f-42cd-aadc-df800f2499d4-config-data\") pod \"92718cf1-7a8f-42cd-aadc-df800f2499d4\" (UID: \"92718cf1-7a8f-42cd-aadc-df800f2499d4\") " Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.240640 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92718cf1-7a8f-42cd-aadc-df800f2499d4-combined-ca-bundle\") pod \"92718cf1-7a8f-42cd-aadc-df800f2499d4\" (UID: \"92718cf1-7a8f-42cd-aadc-df800f2499d4\") " Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.240718 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdh62\" (UniqueName: \"kubernetes.io/projected/92718cf1-7a8f-42cd-aadc-df800f2499d4-kube-api-access-bdh62\") pod \"92718cf1-7a8f-42cd-aadc-df800f2499d4\" (UID: \"92718cf1-7a8f-42cd-aadc-df800f2499d4\") " Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.240753 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bl7hr\" (UniqueName: \"kubernetes.io/projected/78946b93-1c47-42b0-98b4-7a9454ea8dbc-kube-api-access-bl7hr\") pod \"78946b93-1c47-42b0-98b4-7a9454ea8dbc\" (UID: \"78946b93-1c47-42b0-98b4-7a9454ea8dbc\") " Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.240965 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78946b93-1c47-42b0-98b4-7a9454ea8dbc-operator-scripts\") pod \"78946b93-1c47-42b0-98b4-7a9454ea8dbc\" (UID: \"78946b93-1c47-42b0-98b4-7a9454ea8dbc\") " Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.243762 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78946b93-1c47-42b0-98b4-7a9454ea8dbc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "78946b93-1c47-42b0-98b4-7a9454ea8dbc" (UID: "78946b93-1c47-42b0-98b4-7a9454ea8dbc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.248037 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78946b93-1c47-42b0-98b4-7a9454ea8dbc-kube-api-access-bl7hr" (OuterVolumeSpecName: "kube-api-access-bl7hr") pod "78946b93-1c47-42b0-98b4-7a9454ea8dbc" (UID: "78946b93-1c47-42b0-98b4-7a9454ea8dbc"). InnerVolumeSpecName "kube-api-access-bl7hr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.253069 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92718cf1-7a8f-42cd-aadc-df800f2499d4-kube-api-access-bdh62" (OuterVolumeSpecName: "kube-api-access-bdh62") pod "92718cf1-7a8f-42cd-aadc-df800f2499d4" (UID: "92718cf1-7a8f-42cd-aadc-df800f2499d4"). InnerVolumeSpecName "kube-api-access-bdh62". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.287929 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92718cf1-7a8f-42cd-aadc-df800f2499d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "92718cf1-7a8f-42cd-aadc-df800f2499d4" (UID: "92718cf1-7a8f-42cd-aadc-df800f2499d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.311989 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92718cf1-7a8f-42cd-aadc-df800f2499d4-config-data" (OuterVolumeSpecName: "config-data") pod "92718cf1-7a8f-42cd-aadc-df800f2499d4" (UID: "92718cf1-7a8f-42cd-aadc-df800f2499d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.344039 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78946b93-1c47-42b0-98b4-7a9454ea8dbc-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.344084 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92718cf1-7a8f-42cd-aadc-df800f2499d4-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.344100 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92718cf1-7a8f-42cd-aadc-df800f2499d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.344115 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdh62\" (UniqueName: \"kubernetes.io/projected/92718cf1-7a8f-42cd-aadc-df800f2499d4-kube-api-access-bdh62\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.344130 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bl7hr\" (UniqueName: \"kubernetes.io/projected/78946b93-1c47-42b0-98b4-7a9454ea8dbc-kube-api-access-bl7hr\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.514050 4696 generic.go:334] "Generic (PLEG): container finished" podID="7ec50e79-4227-4770-b7bc-7dd5f4178fad" containerID="7c77d3da52a0549309ce8ca707fcc34d55de4d21d9d8040ed15aef36bb950f25" exitCode=0 Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.514109 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" event={"ID":"7ec50e79-4227-4770-b7bc-7dd5f4178fad","Type":"ContainerDied","Data":"7c77d3da52a0549309ce8ca707fcc34d55de4d21d9d8040ed15aef36bb950f25"} Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.514186 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" event={"ID":"7ec50e79-4227-4770-b7bc-7dd5f4178fad","Type":"ContainerStarted","Data":"1ed07f806773bf783d387a27d94bd1dea29d39cc051966a924066d323a143e8d"} Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.515749 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-trn7c" event={"ID":"92718cf1-7a8f-42cd-aadc-df800f2499d4","Type":"ContainerDied","Data":"ade4e3d38b01521263bfd605cfcb97d9bb753d4d0764db1b8aac05dcdd99be0c"} Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.515785 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ade4e3d38b01521263bfd605cfcb97d9bb753d4d0764db1b8aac05dcdd99be0c" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.515790 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-trn7c" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.519239 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-d2jgl" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.519290 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-d2jgl" event={"ID":"78946b93-1c47-42b0-98b4-7a9454ea8dbc","Type":"ContainerDied","Data":"af1e7011478ddfcb87c51458567f070d82bab729a02641c8c2dd986c654dc34c"} Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.519320 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af1e7011478ddfcb87c51458567f070d82bab729a02641c8c2dd986c654dc34c" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.519333 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" podUID="5ce3479f-1bdc-4779-9860-7fe9a59d6e94" containerName="dnsmasq-dns" containerID="cri-o://a95ef0ed856984fc1e9f6ad813bed8db93b34bf18988bcd1f080a5078c1422c2" gracePeriod=10 Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.519443 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.814832 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-bnf6p"] Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.835489 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-2pgpc"] Mar 21 08:47:01 crc kubenswrapper[4696]: E0321 08:47:01.835924 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92718cf1-7a8f-42cd-aadc-df800f2499d4" containerName="keystone-db-sync" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.835940 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="92718cf1-7a8f-42cd-aadc-df800f2499d4" containerName="keystone-db-sync" Mar 21 08:47:01 crc kubenswrapper[4696]: E0321 08:47:01.835970 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78946b93-1c47-42b0-98b4-7a9454ea8dbc" containerName="mariadb-account-create-update" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.835977 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="78946b93-1c47-42b0-98b4-7a9454ea8dbc" containerName="mariadb-account-create-update" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.836149 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="78946b93-1c47-42b0-98b4-7a9454ea8dbc" containerName="mariadb-account-create-update" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.836170 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="92718cf1-7a8f-42cd-aadc-df800f2499d4" containerName="keystone-db-sync" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.836848 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.840975 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-s9frx" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.841218 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.841283 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.841401 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.841520 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.848675 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2pgpc"] Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.851903 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-config-data\") pod \"keystone-bootstrap-2pgpc\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.851977 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-fernet-keys\") pod \"keystone-bootstrap-2pgpc\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.852001 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t92jr\" (UniqueName: \"kubernetes.io/projected/50c2c69e-b049-4b2a-abca-6943a24bdfee-kube-api-access-t92jr\") pod \"keystone-bootstrap-2pgpc\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.852061 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-scripts\") pod \"keystone-bootstrap-2pgpc\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.852089 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-combined-ca-bundle\") pod \"keystone-bootstrap-2pgpc\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.852115 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-credential-keys\") pod \"keystone-bootstrap-2pgpc\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.866991 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-ggnvq"] Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.868561 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.917257 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-ggnvq"] Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.953362 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-dns-svc\") pod \"dnsmasq-dns-847c4cc679-ggnvq\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.953411 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52shz\" (UniqueName: \"kubernetes.io/projected/d621abee-2030-44d8-8052-2db726daac53-kube-api-access-52shz\") pod \"dnsmasq-dns-847c4cc679-ggnvq\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.953433 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-config\") pod \"dnsmasq-dns-847c4cc679-ggnvq\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.953453 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-config-data\") pod \"keystone-bootstrap-2pgpc\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.953497 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-fernet-keys\") pod \"keystone-bootstrap-2pgpc\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.953514 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t92jr\" (UniqueName: \"kubernetes.io/projected/50c2c69e-b049-4b2a-abca-6943a24bdfee-kube-api-access-t92jr\") pod \"keystone-bootstrap-2pgpc\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.953529 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-ggnvq\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.953584 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-scripts\") pod \"keystone-bootstrap-2pgpc\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.953605 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-ggnvq\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.953621 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-combined-ca-bundle\") pod \"keystone-bootstrap-2pgpc\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.953643 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-ggnvq\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.953659 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-credential-keys\") pod \"keystone-bootstrap-2pgpc\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.961235 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-scripts\") pod \"keystone-bootstrap-2pgpc\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.961324 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-config-data\") pod \"keystone-bootstrap-2pgpc\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.962387 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-credential-keys\") pod \"keystone-bootstrap-2pgpc\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.964447 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-fernet-keys\") pod \"keystone-bootstrap-2pgpc\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:01 crc kubenswrapper[4696]: I0321 08:47:01.964637 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-combined-ca-bundle\") pod \"keystone-bootstrap-2pgpc\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.005207 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t92jr\" (UniqueName: \"kubernetes.io/projected/50c2c69e-b049-4b2a-abca-6943a24bdfee-kube-api-access-t92jr\") pod \"keystone-bootstrap-2pgpc\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.093780 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-m25wk"] Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.095224 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-dns-svc\") pod \"dnsmasq-dns-847c4cc679-ggnvq\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.095294 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52shz\" (UniqueName: \"kubernetes.io/projected/d621abee-2030-44d8-8052-2db726daac53-kube-api-access-52shz\") pod \"dnsmasq-dns-847c4cc679-ggnvq\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.095321 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-config\") pod \"dnsmasq-dns-847c4cc679-ggnvq\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.095381 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-ggnvq\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.095461 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-ggnvq\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.095688 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-ggnvq\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.096624 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-ggnvq\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.096629 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-m25wk" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.096916 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-config\") pod \"dnsmasq-dns-847c4cc679-ggnvq\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.102752 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.103222 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.103606 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-9rtmn" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.104068 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-ggnvq\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.104508 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-m25wk"] Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.107507 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-ggnvq\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.109322 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-dns-svc\") pod \"dnsmasq-dns-847c4cc679-ggnvq\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.145679 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52shz\" (UniqueName: \"kubernetes.io/projected/d621abee-2030-44d8-8052-2db726daac53-kube-api-access-52shz\") pod \"dnsmasq-dns-847c4cc679-ggnvq\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.162595 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.208548 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.224211 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-n5tdf"] Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.234555 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.241262 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-92r5d" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.241448 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.241585 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.272393 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-nqvgp"] Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.273505 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-rzcz6"] Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.274305 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rzcz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.274756 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nqvgp" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.277477 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.301137 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpdc5\" (UniqueName: \"kubernetes.io/projected/a70d1c8a-2db3-4ade-af61-fc8225262c18-kube-api-access-dpdc5\") pod \"cinder-db-sync-n5tdf\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.301173 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-config-data\") pod \"cinder-db-sync-n5tdf\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.301237 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-combined-ca-bundle\") pod \"cinder-db-sync-n5tdf\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.301269 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4ac0955-0d2a-4081-b82b-32dd288de552-combined-ca-bundle\") pod \"neutron-db-sync-m25wk\" (UID: \"c4ac0955-0d2a-4081-b82b-32dd288de552\") " pod="openstack/neutron-db-sync-m25wk" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.301290 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c4ac0955-0d2a-4081-b82b-32dd288de552-config\") pod \"neutron-db-sync-m25wk\" (UID: \"c4ac0955-0d2a-4081-b82b-32dd288de552\") " pod="openstack/neutron-db-sync-m25wk" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.301308 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-scripts\") pod \"cinder-db-sync-n5tdf\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.301332 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-db-sync-config-data\") pod \"cinder-db-sync-n5tdf\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.301377 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a70d1c8a-2db3-4ade-af61-fc8225262c18-etc-machine-id\") pod \"cinder-db-sync-n5tdf\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.301428 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrw9c\" (UniqueName: \"kubernetes.io/projected/c4ac0955-0d2a-4081-b82b-32dd288de552-kube-api-access-xrw9c\") pod \"neutron-db-sync-m25wk\" (UID: \"c4ac0955-0d2a-4081-b82b-32dd288de552\") " pod="openstack/neutron-db-sync-m25wk" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.309788 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.310045 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.310209 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-6hxd9" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.310345 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-f2wb5" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.310457 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nqvgp"] Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.360240 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-n5tdf"] Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.396415 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-rzcz6"] Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.405447 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80364601-ba09-4192-a176-e04ad8d45506-config-data\") pod \"placement-db-sync-nqvgp\" (UID: \"80364601-ba09-4192-a176-e04ad8d45506\") " pod="openstack/placement-db-sync-nqvgp" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.405489 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80364601-ba09-4192-a176-e04ad8d45506-combined-ca-bundle\") pod \"placement-db-sync-nqvgp\" (UID: \"80364601-ba09-4192-a176-e04ad8d45506\") " pod="openstack/placement-db-sync-nqvgp" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.405601 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrw9c\" (UniqueName: \"kubernetes.io/projected/c4ac0955-0d2a-4081-b82b-32dd288de552-kube-api-access-xrw9c\") pod \"neutron-db-sync-m25wk\" (UID: \"c4ac0955-0d2a-4081-b82b-32dd288de552\") " pod="openstack/neutron-db-sync-m25wk" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.405665 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpdc5\" (UniqueName: \"kubernetes.io/projected/a70d1c8a-2db3-4ade-af61-fc8225262c18-kube-api-access-dpdc5\") pod \"cinder-db-sync-n5tdf\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.405684 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-config-data\") pod \"cinder-db-sync-n5tdf\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.405732 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4381bdfb-820a-42cc-9a0f-0ca3e6f34449-db-sync-config-data\") pod \"barbican-db-sync-rzcz6\" (UID: \"4381bdfb-820a-42cc-9a0f-0ca3e6f34449\") " pod="openstack/barbican-db-sync-rzcz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.405759 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stvs7\" (UniqueName: \"kubernetes.io/projected/4381bdfb-820a-42cc-9a0f-0ca3e6f34449-kube-api-access-stvs7\") pod \"barbican-db-sync-rzcz6\" (UID: \"4381bdfb-820a-42cc-9a0f-0ca3e6f34449\") " pod="openstack/barbican-db-sync-rzcz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.406170 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4381bdfb-820a-42cc-9a0f-0ca3e6f34449-combined-ca-bundle\") pod \"barbican-db-sync-rzcz6\" (UID: \"4381bdfb-820a-42cc-9a0f-0ca3e6f34449\") " pod="openstack/barbican-db-sync-rzcz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.406193 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80364601-ba09-4192-a176-e04ad8d45506-logs\") pod \"placement-db-sync-nqvgp\" (UID: \"80364601-ba09-4192-a176-e04ad8d45506\") " pod="openstack/placement-db-sync-nqvgp" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.406230 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80364601-ba09-4192-a176-e04ad8d45506-scripts\") pod \"placement-db-sync-nqvgp\" (UID: \"80364601-ba09-4192-a176-e04ad8d45506\") " pod="openstack/placement-db-sync-nqvgp" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.406253 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-combined-ca-bundle\") pod \"cinder-db-sync-n5tdf\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.406307 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4ac0955-0d2a-4081-b82b-32dd288de552-combined-ca-bundle\") pod \"neutron-db-sync-m25wk\" (UID: \"c4ac0955-0d2a-4081-b82b-32dd288de552\") " pod="openstack/neutron-db-sync-m25wk" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.406336 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c4ac0955-0d2a-4081-b82b-32dd288de552-config\") pod \"neutron-db-sync-m25wk\" (UID: \"c4ac0955-0d2a-4081-b82b-32dd288de552\") " pod="openstack/neutron-db-sync-m25wk" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.406360 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-scripts\") pod \"cinder-db-sync-n5tdf\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.406385 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-db-sync-config-data\") pod \"cinder-db-sync-n5tdf\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.406480 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a70d1c8a-2db3-4ade-af61-fc8225262c18-etc-machine-id\") pod \"cinder-db-sync-n5tdf\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.406501 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trrnn\" (UniqueName: \"kubernetes.io/projected/80364601-ba09-4192-a176-e04ad8d45506-kube-api-access-trrnn\") pod \"placement-db-sync-nqvgp\" (UID: \"80364601-ba09-4192-a176-e04ad8d45506\") " pod="openstack/placement-db-sync-nqvgp" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.411378 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-config-data\") pod \"cinder-db-sync-n5tdf\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.427739 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a70d1c8a-2db3-4ade-af61-fc8225262c18-etc-machine-id\") pod \"cinder-db-sync-n5tdf\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.434422 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4ac0955-0d2a-4081-b82b-32dd288de552-combined-ca-bundle\") pod \"neutron-db-sync-m25wk\" (UID: \"c4ac0955-0d2a-4081-b82b-32dd288de552\") " pod="openstack/neutron-db-sync-m25wk" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.435690 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-combined-ca-bundle\") pod \"cinder-db-sync-n5tdf\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.475767 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrw9c\" (UniqueName: \"kubernetes.io/projected/c4ac0955-0d2a-4081-b82b-32dd288de552-kube-api-access-xrw9c\") pod \"neutron-db-sync-m25wk\" (UID: \"c4ac0955-0d2a-4081-b82b-32dd288de552\") " pod="openstack/neutron-db-sync-m25wk" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.476365 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c4ac0955-0d2a-4081-b82b-32dd288de552-config\") pod \"neutron-db-sync-m25wk\" (UID: \"c4ac0955-0d2a-4081-b82b-32dd288de552\") " pod="openstack/neutron-db-sync-m25wk" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.479057 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-db-sync-vgsz6"] Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.480622 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-vgsz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.489135 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.489264 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.491347 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-ff6v6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.497922 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.498604 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-db-sync-config-data\") pod \"cinder-db-sync-n5tdf\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.511132 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trrnn\" (UniqueName: \"kubernetes.io/projected/80364601-ba09-4192-a176-e04ad8d45506-kube-api-access-trrnn\") pod \"placement-db-sync-nqvgp\" (UID: \"80364601-ba09-4192-a176-e04ad8d45506\") " pod="openstack/placement-db-sync-nqvgp" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.511180 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80364601-ba09-4192-a176-e04ad8d45506-config-data\") pod \"placement-db-sync-nqvgp\" (UID: \"80364601-ba09-4192-a176-e04ad8d45506\") " pod="openstack/placement-db-sync-nqvgp" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.511219 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80364601-ba09-4192-a176-e04ad8d45506-combined-ca-bundle\") pod \"placement-db-sync-nqvgp\" (UID: \"80364601-ba09-4192-a176-e04ad8d45506\") " pod="openstack/placement-db-sync-nqvgp" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.511326 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4381bdfb-820a-42cc-9a0f-0ca3e6f34449-db-sync-config-data\") pod \"barbican-db-sync-rzcz6\" (UID: \"4381bdfb-820a-42cc-9a0f-0ca3e6f34449\") " pod="openstack/barbican-db-sync-rzcz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.511377 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stvs7\" (UniqueName: \"kubernetes.io/projected/4381bdfb-820a-42cc-9a0f-0ca3e6f34449-kube-api-access-stvs7\") pod \"barbican-db-sync-rzcz6\" (UID: \"4381bdfb-820a-42cc-9a0f-0ca3e6f34449\") " pod="openstack/barbican-db-sync-rzcz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.511405 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4381bdfb-820a-42cc-9a0f-0ca3e6f34449-combined-ca-bundle\") pod \"barbican-db-sync-rzcz6\" (UID: \"4381bdfb-820a-42cc-9a0f-0ca3e6f34449\") " pod="openstack/barbican-db-sync-rzcz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.511446 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80364601-ba09-4192-a176-e04ad8d45506-logs\") pod \"placement-db-sync-nqvgp\" (UID: \"80364601-ba09-4192-a176-e04ad8d45506\") " pod="openstack/placement-db-sync-nqvgp" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.511472 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80364601-ba09-4192-a176-e04ad8d45506-scripts\") pod \"placement-db-sync-nqvgp\" (UID: \"80364601-ba09-4192-a176-e04ad8d45506\") " pod="openstack/placement-db-sync-nqvgp" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.521173 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80364601-ba09-4192-a176-e04ad8d45506-logs\") pod \"placement-db-sync-nqvgp\" (UID: \"80364601-ba09-4192-a176-e04ad8d45506\") " pod="openstack/placement-db-sync-nqvgp" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.544283 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-ggnvq"] Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.583258 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-m25wk" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.585977 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-scripts\") pod \"cinder-db-sync-n5tdf\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.586432 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80364601-ba09-4192-a176-e04ad8d45506-config-data\") pod \"placement-db-sync-nqvgp\" (UID: \"80364601-ba09-4192-a176-e04ad8d45506\") " pod="openstack/placement-db-sync-nqvgp" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.599416 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpdc5\" (UniqueName: \"kubernetes.io/projected/a70d1c8a-2db3-4ade-af61-fc8225262c18-kube-api-access-dpdc5\") pod \"cinder-db-sync-n5tdf\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.609128 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stvs7\" (UniqueName: \"kubernetes.io/projected/4381bdfb-820a-42cc-9a0f-0ca3e6f34449-kube-api-access-stvs7\") pod \"barbican-db-sync-rzcz6\" (UID: \"4381bdfb-820a-42cc-9a0f-0ca3e6f34449\") " pod="openstack/barbican-db-sync-rzcz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.609302 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80364601-ba09-4192-a176-e04ad8d45506-scripts\") pod \"placement-db-sync-nqvgp\" (UID: \"80364601-ba09-4192-a176-e04ad8d45506\") " pod="openstack/placement-db-sync-nqvgp" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.612152 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80364601-ba09-4192-a176-e04ad8d45506-combined-ca-bundle\") pod \"placement-db-sync-nqvgp\" (UID: \"80364601-ba09-4192-a176-e04ad8d45506\") " pod="openstack/placement-db-sync-nqvgp" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.612621 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mv7dw\" (UniqueName: \"kubernetes.io/projected/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-kube-api-access-mv7dw\") pod \"cloudkitty-db-sync-vgsz6\" (UID: \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\") " pod="openstack/cloudkitty-db-sync-vgsz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.612727 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-combined-ca-bundle\") pod \"cloudkitty-db-sync-vgsz6\" (UID: \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\") " pod="openstack/cloudkitty-db-sync-vgsz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.612792 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-config-data\") pod \"cloudkitty-db-sync-vgsz6\" (UID: \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\") " pod="openstack/cloudkitty-db-sync-vgsz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.612825 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-scripts\") pod \"cloudkitty-db-sync-vgsz6\" (UID: \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\") " pod="openstack/cloudkitty-db-sync-vgsz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.612849 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-certs\") pod \"cloudkitty-db-sync-vgsz6\" (UID: \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\") " pod="openstack/cloudkitty-db-sync-vgsz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.614100 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4381bdfb-820a-42cc-9a0f-0ca3e6f34449-combined-ca-bundle\") pod \"barbican-db-sync-rzcz6\" (UID: \"4381bdfb-820a-42cc-9a0f-0ca3e6f34449\") " pod="openstack/barbican-db-sync-rzcz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.617256 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4381bdfb-820a-42cc-9a0f-0ca3e6f34449-db-sync-config-data\") pod \"barbican-db-sync-rzcz6\" (UID: \"4381bdfb-820a-42cc-9a0f-0ca3e6f34449\") " pod="openstack/barbican-db-sync-rzcz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.617677 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trrnn\" (UniqueName: \"kubernetes.io/projected/80364601-ba09-4192-a176-e04ad8d45506-kube-api-access-trrnn\") pod \"placement-db-sync-nqvgp\" (UID: \"80364601-ba09-4192-a176-e04ad8d45506\") " pod="openstack/placement-db-sync-nqvgp" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.624257 4696 generic.go:334] "Generic (PLEG): container finished" podID="5ce3479f-1bdc-4779-9860-7fe9a59d6e94" containerID="a95ef0ed856984fc1e9f6ad813bed8db93b34bf18988bcd1f080a5078c1422c2" exitCode=0 Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.638998 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nqvgp" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.665068 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.665872 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-vgsz6"] Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.665904 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" event={"ID":"7ec50e79-4227-4770-b7bc-7dd5f4178fad","Type":"ContainerStarted","Data":"633a094e259945b090fda18510d72f11afe12b0a1f579282734a24c2e455b0ee"} Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.665939 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" event={"ID":"5ce3479f-1bdc-4779-9860-7fe9a59d6e94","Type":"ContainerDied","Data":"a95ef0ed856984fc1e9f6ad813bed8db93b34bf18988bcd1f080a5078c1422c2"} Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.665956 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" event={"ID":"5ce3479f-1bdc-4779-9860-7fe9a59d6e94","Type":"ContainerDied","Data":"e8c610b780fc56bec96452b55b4cad0181baff6aa276397f696edc73fc763d78"} Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.665967 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8c610b780fc56bec96452b55b4cad0181baff6aa276397f696edc73fc763d78" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.665977 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-mddgz"] Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.666471 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.674240 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.698431 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-mddgz"] Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.716203 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-combined-ca-bundle\") pod \"cloudkitty-db-sync-vgsz6\" (UID: \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\") " pod="openstack/cloudkitty-db-sync-vgsz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.716359 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-config-data\") pod \"cloudkitty-db-sync-vgsz6\" (UID: \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\") " pod="openstack/cloudkitty-db-sync-vgsz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.716383 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-scripts\") pod \"cloudkitty-db-sync-vgsz6\" (UID: \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\") " pod="openstack/cloudkitty-db-sync-vgsz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.716413 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-certs\") pod \"cloudkitty-db-sync-vgsz6\" (UID: \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\") " pod="openstack/cloudkitty-db-sync-vgsz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.716478 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mv7dw\" (UniqueName: \"kubernetes.io/projected/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-kube-api-access-mv7dw\") pod \"cloudkitty-db-sync-vgsz6\" (UID: \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\") " pod="openstack/cloudkitty-db-sync-vgsz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.717732 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.727415 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rzcz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.746258 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-scripts\") pod \"cloudkitty-db-sync-vgsz6\" (UID: \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\") " pod="openstack/cloudkitty-db-sync-vgsz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.756715 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mv7dw\" (UniqueName: \"kubernetes.io/projected/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-kube-api-access-mv7dw\") pod \"cloudkitty-db-sync-vgsz6\" (UID: \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\") " pod="openstack/cloudkitty-db-sync-vgsz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.775231 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-combined-ca-bundle\") pod \"cloudkitty-db-sync-vgsz6\" (UID: \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\") " pod="openstack/cloudkitty-db-sync-vgsz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.777681 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-config-data\") pod \"cloudkitty-db-sync-vgsz6\" (UID: \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\") " pod="openstack/cloudkitty-db-sync-vgsz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.779209 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-certs\") pod \"cloudkitty-db-sync-vgsz6\" (UID: \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\") " pod="openstack/cloudkitty-db-sync-vgsz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.820471 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-dns-swift-storage-0\") pod \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.820538 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-ovsdbserver-nb\") pod \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.820624 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzh27\" (UniqueName: \"kubernetes.io/projected/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-kube-api-access-dzh27\") pod \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.820745 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-ovsdbserver-sb\") pod \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.820779 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-dns-svc\") pod \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.820854 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-config\") pod \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\" (UID: \"5ce3479f-1bdc-4779-9860-7fe9a59d6e94\") " Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.821312 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frckn\" (UniqueName: \"kubernetes.io/projected/64b14a78-860a-48c3-ae69-16cc36c77be8-kube-api-access-frckn\") pod \"dnsmasq-dns-785d8bcb8c-mddgz\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.821410 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-mddgz\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.821471 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-mddgz\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.821511 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-mddgz\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.821588 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-config\") pod \"dnsmasq-dns-785d8bcb8c-mddgz\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.821639 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-mddgz\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.853972 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-kube-api-access-dzh27" (OuterVolumeSpecName: "kube-api-access-dzh27") pod "5ce3479f-1bdc-4779-9860-7fe9a59d6e94" (UID: "5ce3479f-1bdc-4779-9860-7fe9a59d6e94"). InnerVolumeSpecName "kube-api-access-dzh27". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.881728 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" podStartSLOduration=2.881709333 podStartE2EDuration="2.881709333s" podCreationTimestamp="2026-03-21 08:47:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:47:02.615932346 +0000 UTC m=+1156.736813059" watchObservedRunningTime="2026-03-21 08:47:02.881709333 +0000 UTC m=+1157.002590046" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.926005 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-mddgz\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.933628 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-mddgz\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.933756 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-config\") pod \"dnsmasq-dns-785d8bcb8c-mddgz\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.933868 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-mddgz\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.934009 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frckn\" (UniqueName: \"kubernetes.io/projected/64b14a78-860a-48c3-ae69-16cc36c77be8-kube-api-access-frckn\") pod \"dnsmasq-dns-785d8bcb8c-mddgz\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.934128 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-mddgz\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.934236 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzh27\" (UniqueName: \"kubernetes.io/projected/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-kube-api-access-dzh27\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.940508 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-mddgz\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.946884 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-config\") pod \"dnsmasq-dns-785d8bcb8c-mddgz\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.947291 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-mddgz\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.947473 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-mddgz\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.948580 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-mddgz\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.978935 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frckn\" (UniqueName: \"kubernetes.io/projected/64b14a78-860a-48c3-ae69-16cc36c77be8-kube-api-access-frckn\") pod \"dnsmasq-dns-785d8bcb8c-mddgz\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.989789 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:47:02 crc kubenswrapper[4696]: E0321 08:47:02.990289 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ce3479f-1bdc-4779-9860-7fe9a59d6e94" containerName="dnsmasq-dns" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.990302 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ce3479f-1bdc-4779-9860-7fe9a59d6e94" containerName="dnsmasq-dns" Mar 21 08:47:02 crc kubenswrapper[4696]: E0321 08:47:02.990321 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ce3479f-1bdc-4779-9860-7fe9a59d6e94" containerName="init" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.990328 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ce3479f-1bdc-4779-9860-7fe9a59d6e94" containerName="init" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.990387 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5ce3479f-1bdc-4779-9860-7fe9a59d6e94" (UID: "5ce3479f-1bdc-4779-9860-7fe9a59d6e94"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.990488 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ce3479f-1bdc-4779-9860-7fe9a59d6e94" containerName="dnsmasq-dns" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.992150 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.993384 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-vgsz6" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.995460 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 21 08:47:02 crc kubenswrapper[4696]: I0321 08:47:02.996686 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.007158 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-config" (OuterVolumeSpecName: "config") pod "5ce3479f-1bdc-4779-9860-7fe9a59d6e94" (UID: "5ce3479f-1bdc-4779-9860-7fe9a59d6e94"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.034301 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.035738 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5ce3479f-1bdc-4779-9860-7fe9a59d6e94" (UID: "5ce3479f-1bdc-4779-9860-7fe9a59d6e94"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.040270 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.043299 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbd625f3-5fcc-48f5-9022-df08fdc01887-log-httpd\") pod \"ceilometer-0\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.043373 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.043624 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.044086 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbd625f3-5fcc-48f5-9022-df08fdc01887-run-httpd\") pod \"ceilometer-0\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.045156 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-scripts\") pod \"ceilometer-0\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.045628 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4crtt\" (UniqueName: \"kubernetes.io/projected/dbd625f3-5fcc-48f5-9022-df08fdc01887-kube-api-access-4crtt\") pod \"ceilometer-0\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.045739 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-config-data\") pod \"ceilometer-0\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.046567 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.046591 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.046601 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.063223 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5ce3479f-1bdc-4779-9860-7fe9a59d6e94" (UID: "5ce3479f-1bdc-4779-9860-7fe9a59d6e94"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.065251 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5ce3479f-1bdc-4779-9860-7fe9a59d6e94" (UID: "5ce3479f-1bdc-4779-9860-7fe9a59d6e94"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.067186 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.071517 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.074369 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-f5cmp" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.078778 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.080775 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.088094 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.105802 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.108092 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.115987 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.124508 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.148342 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01e35b12-5b6a-4cdd-941f-b2efb0aca029-scripts\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.148394 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fzxp\" (UniqueName: \"kubernetes.io/projected/2787955f-dbec-430c-91b1-7554a2b20b23-kube-api-access-2fzxp\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.148431 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2787955f-dbec-430c-91b1-7554a2b20b23-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.148503 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2787955f-dbec-430c-91b1-7554a2b20b23-scripts\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.148569 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.148622 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01e35b12-5b6a-4cdd-941f-b2efb0aca029-logs\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.148659 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-scripts\") pod \"ceilometer-0\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.148681 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2787955f-dbec-430c-91b1-7554a2b20b23-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.148723 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4crtt\" (UniqueName: \"kubernetes.io/projected/dbd625f3-5fcc-48f5-9022-df08fdc01887-kube-api-access-4crtt\") pod \"ceilometer-0\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.148749 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r45cw\" (UniqueName: \"kubernetes.io/projected/01e35b12-5b6a-4cdd-941f-b2efb0aca029-kube-api-access-r45cw\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.148809 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-config-data\") pod \"ceilometer-0\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.148853 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/01e35b12-5b6a-4cdd-941f-b2efb0aca029-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.148881 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbd625f3-5fcc-48f5-9022-df08fdc01887-log-httpd\") pod \"ceilometer-0\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.148924 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.148959 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2787955f-dbec-430c-91b1-7554a2b20b23-config-data\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.148987 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01e35b12-5b6a-4cdd-941f-b2efb0aca029-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.149016 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.149056 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01e35b12-5b6a-4cdd-941f-b2efb0aca029-config-data\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.149083 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.149126 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2787955f-dbec-430c-91b1-7554a2b20b23-logs\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.149152 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbd625f3-5fcc-48f5-9022-df08fdc01887-run-httpd\") pod \"ceilometer-0\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.149220 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.149247 4696 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ce3479f-1bdc-4779-9860-7fe9a59d6e94-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.149698 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbd625f3-5fcc-48f5-9022-df08fdc01887-run-httpd\") pod \"ceilometer-0\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.149895 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbd625f3-5fcc-48f5-9022-df08fdc01887-log-httpd\") pod \"ceilometer-0\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.158457 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-scripts\") pod \"ceilometer-0\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.164805 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.165899 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.166356 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-config-data\") pod \"ceilometer-0\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.168066 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4crtt\" (UniqueName: \"kubernetes.io/projected/dbd625f3-5fcc-48f5-9022-df08fdc01887-kube-api-access-4crtt\") pod \"ceilometer-0\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.250442 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01e35b12-5b6a-4cdd-941f-b2efb0aca029-logs\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.250795 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2787955f-dbec-430c-91b1-7554a2b20b23-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.250844 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r45cw\" (UniqueName: \"kubernetes.io/projected/01e35b12-5b6a-4cdd-941f-b2efb0aca029-kube-api-access-r45cw\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.251067 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/01e35b12-5b6a-4cdd-941f-b2efb0aca029-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.251111 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2787955f-dbec-430c-91b1-7554a2b20b23-config-data\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.251131 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01e35b12-5b6a-4cdd-941f-b2efb0aca029-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.251152 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.251180 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01e35b12-5b6a-4cdd-941f-b2efb0aca029-config-data\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.251212 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2787955f-dbec-430c-91b1-7554a2b20b23-logs\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.251237 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01e35b12-5b6a-4cdd-941f-b2efb0aca029-scripts\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.251257 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fzxp\" (UniqueName: \"kubernetes.io/projected/2787955f-dbec-430c-91b1-7554a2b20b23-kube-api-access-2fzxp\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.251281 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2787955f-dbec-430c-91b1-7554a2b20b23-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.251329 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2787955f-dbec-430c-91b1-7554a2b20b23-scripts\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.251386 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.252922 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/01e35b12-5b6a-4cdd-941f-b2efb0aca029-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.253363 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01e35b12-5b6a-4cdd-941f-b2efb0aca029-logs\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.253393 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2787955f-dbec-430c-91b1-7554a2b20b23-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.253646 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2787955f-dbec-430c-91b1-7554a2b20b23-logs\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.263283 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01e35b12-5b6a-4cdd-941f-b2efb0aca029-config-data\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.264457 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01e35b12-5b6a-4cdd-941f-b2efb0aca029-scripts\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.280573 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.280613 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/36feb21b6af5aad5e90205d24e3f1d239a0e9aaaadf2ca8cf7f1809b63388e1a/globalmount\"" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.287104 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r45cw\" (UniqueName: \"kubernetes.io/projected/01e35b12-5b6a-4cdd-941f-b2efb0aca029-kube-api-access-r45cw\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.289642 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fzxp\" (UniqueName: \"kubernetes.io/projected/2787955f-dbec-430c-91b1-7554a2b20b23-kube-api-access-2fzxp\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.295731 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2787955f-dbec-430c-91b1-7554a2b20b23-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.299155 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01e35b12-5b6a-4cdd-941f-b2efb0aca029-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.299762 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2787955f-dbec-430c-91b1-7554a2b20b23-scripts\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.300364 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.300388 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0bb60238debc80c6ed39528f98bb598b47b2c7acc3e22f58b887f712b724eea4/globalmount\"" pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.309580 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2787955f-dbec-430c-91b1-7554a2b20b23-config-data\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.348976 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") pod \"glance-default-external-api-0\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.359756 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") pod \"glance-default-internal-api-0\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.474912 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.488365 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.532551 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2pgpc"] Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.596538 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.659046 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-v6pnl" Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.659434 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2pgpc" event={"ID":"50c2c69e-b049-4b2a-abca-6943a24bdfee","Type":"ContainerStarted","Data":"8c1e7a3aba219c87466ffc5f362517623ce0a2ce62f094e2d24dc64a00d53b5f"} Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.659845 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" podUID="7ec50e79-4227-4770-b7bc-7dd5f4178fad" containerName="dnsmasq-dns" containerID="cri-o://633a094e259945b090fda18510d72f11afe12b0a1f579282734a24c2e455b0ee" gracePeriod=10 Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.751884 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nqvgp"] Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.790049 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-ggnvq"] Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.811584 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-v6pnl"] Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.839190 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-v6pnl"] Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.919324 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-m25wk"] Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.931206 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-rzcz6"] Mar 21 08:47:03 crc kubenswrapper[4696]: W0321 08:47:03.956445 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4381bdfb_820a_42cc_9a0f_0ca3e6f34449.slice/crio-38e1b22cb3a6ff8a04a90b85b6e070a3e96f95b86b0eeb9dbe2d46b4b0773b3d WatchSource:0}: Error finding container 38e1b22cb3a6ff8a04a90b85b6e070a3e96f95b86b0eeb9dbe2d46b4b0773b3d: Status 404 returned error can't find the container with id 38e1b22cb3a6ff8a04a90b85b6e070a3e96f95b86b0eeb9dbe2d46b4b0773b3d Mar 21 08:47:03 crc kubenswrapper[4696]: I0321 08:47:03.959794 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-n5tdf"] Mar 21 08:47:04 crc kubenswrapper[4696]: W0321 08:47:04.312553 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64b14a78_860a_48c3_ae69_16cc36c77be8.slice/crio-037087306e9eaf6637464e8aaa5a52f1a30f4bd4e5326e7b9bb91a6a781071e8 WatchSource:0}: Error finding container 037087306e9eaf6637464e8aaa5a52f1a30f4bd4e5326e7b9bb91a6a781071e8: Status 404 returned error can't find the container with id 037087306e9eaf6637464e8aaa5a52f1a30f4bd4e5326e7b9bb91a6a781071e8 Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.319587 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-mddgz"] Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.440456 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-vgsz6"] Mar 21 08:47:04 crc kubenswrapper[4696]: W0321 08:47:04.449993 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3fc5362_53e5_4c35_be23_49dc3dd1fcba.slice/crio-bbfd9c99355e5907665a8e207c943040f2efbaba94ae5f6bf00a922e2aad9fdd WatchSource:0}: Error finding container bbfd9c99355e5907665a8e207c943040f2efbaba94ae5f6bf00a922e2aad9fdd: Status 404 returned error can't find the container with id bbfd9c99355e5907665a8e207c943040f2efbaba94ae5f6bf00a922e2aad9fdd Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.488420 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.517886 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.622612 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ce3479f-1bdc-4779-9860-7fe9a59d6e94" path="/var/lib/kubelet/pods/5ce3479f-1bdc-4779-9860-7fe9a59d6e94/volumes" Mar 21 08:47:04 crc kubenswrapper[4696]: W0321 08:47:04.626252 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2787955f_dbec_430c_91b1_7554a2b20b23.slice/crio-bc0f472e71fa433353954b6d8102751a93e769117f3aa7371216a6fc2fc8ed26 WatchSource:0}: Error finding container bc0f472e71fa433353954b6d8102751a93e769117f3aa7371216a6fc2fc8ed26: Status 404 returned error can't find the container with id bc0f472e71fa433353954b6d8102751a93e769117f3aa7371216a6fc2fc8ed26 Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.628546 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.674953 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.689846 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-dns-svc\") pod \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.690066 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-dns-swift-storage-0\") pod \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.690192 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-ovsdbserver-nb\") pod \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.690230 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-ovsdbserver-sb\") pod \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.690673 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-config\") pod \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.690748 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjd98\" (UniqueName: \"kubernetes.io/projected/7ec50e79-4227-4770-b7bc-7dd5f4178fad-kube-api-access-fjd98\") pod \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\" (UID: \"7ec50e79-4227-4770-b7bc-7dd5f4178fad\") " Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.736814 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nqvgp" event={"ID":"80364601-ba09-4192-a176-e04ad8d45506","Type":"ContainerStarted","Data":"34f92239a190f120cdae216575bf329d8b3b5c80eb8a4139c0367ba3daa6b441"} Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.737120 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ec50e79-4227-4770-b7bc-7dd5f4178fad-kube-api-access-fjd98" (OuterVolumeSpecName: "kube-api-access-fjd98") pod "7ec50e79-4227-4770-b7bc-7dd5f4178fad" (UID: "7ec50e79-4227-4770-b7bc-7dd5f4178fad"). InnerVolumeSpecName "kube-api-access-fjd98". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.766631 4696 generic.go:334] "Generic (PLEG): container finished" podID="d621abee-2030-44d8-8052-2db726daac53" containerID="78e4f6e2baa3be1cd5da901303d3358ab33a3fad97d133fce46abf24fd584fd0" exitCode=0 Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.766712 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" event={"ID":"d621abee-2030-44d8-8052-2db726daac53","Type":"ContainerDied","Data":"78e4f6e2baa3be1cd5da901303d3358ab33a3fad97d133fce46abf24fd584fd0"} Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.766744 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" event={"ID":"d621abee-2030-44d8-8052-2db726daac53","Type":"ContainerStarted","Data":"49a26d9ada18a6a9d18f2a4b8fa8f44791c5c82dc9e408f2367f352a3d7c8ff1"} Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.793377 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjd98\" (UniqueName: \"kubernetes.io/projected/7ec50e79-4227-4770-b7bc-7dd5f4178fad-kube-api-access-fjd98\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.816702 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rzcz6" event={"ID":"4381bdfb-820a-42cc-9a0f-0ca3e6f34449","Type":"ContainerStarted","Data":"38e1b22cb3a6ff8a04a90b85b6e070a3e96f95b86b0eeb9dbe2d46b4b0773b3d"} Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.829832 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.830251 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbd625f3-5fcc-48f5-9022-df08fdc01887","Type":"ContainerStarted","Data":"e0430d561aa504fb166fa0c66a1816a40fd58d7f009269cec7b22c7098f0ce92"} Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.884890 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2pgpc" event={"ID":"50c2c69e-b049-4b2a-abca-6943a24bdfee","Type":"ContainerStarted","Data":"027b0a845d9b25edaa4aec6a90d1a489fea68982c21b4e5196545686f124576a"} Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.911138 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7ec50e79-4227-4770-b7bc-7dd5f4178fad" (UID: "7ec50e79-4227-4770-b7bc-7dd5f4178fad"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.922479 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-config" (OuterVolumeSpecName: "config") pod "7ec50e79-4227-4770-b7bc-7dd5f4178fad" (UID: "7ec50e79-4227-4770-b7bc-7dd5f4178fad"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.931927 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-m25wk" event={"ID":"c4ac0955-0d2a-4081-b82b-32dd288de552","Type":"ContainerStarted","Data":"f07398703df0547d265b0236bc994a5f3e2b860e2c22130759ce02432acf9cdf"} Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.931977 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-m25wk" event={"ID":"c4ac0955-0d2a-4081-b82b-32dd288de552","Type":"ContainerStarted","Data":"f0d6a379964db7a6dd417f343874b928e2d85d3a88ed27102b05589c7808f438"} Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.939385 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7ec50e79-4227-4770-b7bc-7dd5f4178fad" (UID: "7ec50e79-4227-4770-b7bc-7dd5f4178fad"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.948934 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.956723 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-2pgpc" podStartSLOduration=3.956706347 podStartE2EDuration="3.956706347s" podCreationTimestamp="2026-03-21 08:47:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:47:04.955790141 +0000 UTC m=+1159.076670854" watchObservedRunningTime="2026-03-21 08:47:04.956706347 +0000 UTC m=+1159.077587050" Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.983850 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7ec50e79-4227-4770-b7bc-7dd5f4178fad" (UID: "7ec50e79-4227-4770-b7bc-7dd5f4178fad"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.983911 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7ec50e79-4227-4770-b7bc-7dd5f4178fad" (UID: "7ec50e79-4227-4770-b7bc-7dd5f4178fad"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.987189 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" event={"ID":"64b14a78-860a-48c3-ae69-16cc36c77be8","Type":"ContainerStarted","Data":"3ea87006f13706ba3d8ac0954309bbbc6e34e6fc1350bd3952ec2ae56578a677"} Mar 21 08:47:04 crc kubenswrapper[4696]: I0321 08:47:04.987237 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" event={"ID":"64b14a78-860a-48c3-ae69-16cc36c77be8","Type":"ContainerStarted","Data":"037087306e9eaf6637464e8aaa5a52f1a30f4bd4e5326e7b9bb91a6a781071e8"} Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.008495 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.008521 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.008532 4696 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.008542 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.008551 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ec50e79-4227-4770-b7bc-7dd5f4178fad-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.054915 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-m25wk" podStartSLOduration=3.054886066 podStartE2EDuration="3.054886066s" podCreationTimestamp="2026-03-21 08:47:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:47:05.010316002 +0000 UTC m=+1159.131196715" watchObservedRunningTime="2026-03-21 08:47:05.054886066 +0000 UTC m=+1159.175766779" Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.111047 4696 generic.go:334] "Generic (PLEG): container finished" podID="7ec50e79-4227-4770-b7bc-7dd5f4178fad" containerID="633a094e259945b090fda18510d72f11afe12b0a1f579282734a24c2e455b0ee" exitCode=0 Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.111145 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" event={"ID":"7ec50e79-4227-4770-b7bc-7dd5f4178fad","Type":"ContainerDied","Data":"633a094e259945b090fda18510d72f11afe12b0a1f579282734a24c2e455b0ee"} Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.111178 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" event={"ID":"7ec50e79-4227-4770-b7bc-7dd5f4178fad","Type":"ContainerDied","Data":"1ed07f806773bf783d387a27d94bd1dea29d39cc051966a924066d323a143e8d"} Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.111198 4696 scope.go:117] "RemoveContainer" containerID="633a094e259945b090fda18510d72f11afe12b0a1f579282734a24c2e455b0ee" Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.111340 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-bnf6p" Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.130167 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-n5tdf" event={"ID":"a70d1c8a-2db3-4ade-af61-fc8225262c18","Type":"ContainerStarted","Data":"30a2231543f7973200c0970b434c4ad623190a2282d35287b73a1da3e28caf22"} Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.137736 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-vgsz6" event={"ID":"d3fc5362-53e5-4c35-be23-49dc3dd1fcba","Type":"ContainerStarted","Data":"bbfd9c99355e5907665a8e207c943040f2efbaba94ae5f6bf00a922e2aad9fdd"} Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.145687 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.149098 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2787955f-dbec-430c-91b1-7554a2b20b23","Type":"ContainerStarted","Data":"bc0f472e71fa433353954b6d8102751a93e769117f3aa7371216a6fc2fc8ed26"} Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.316810 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-bnf6p"] Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.375482 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-bnf6p"] Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.571636 4696 scope.go:117] "RemoveContainer" containerID="7c77d3da52a0549309ce8ca707fcc34d55de4d21d9d8040ed15aef36bb950f25" Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.885806 4696 scope.go:117] "RemoveContainer" containerID="633a094e259945b090fda18510d72f11afe12b0a1f579282734a24c2e455b0ee" Mar 21 08:47:05 crc kubenswrapper[4696]: E0321 08:47:05.886898 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"633a094e259945b090fda18510d72f11afe12b0a1f579282734a24c2e455b0ee\": container with ID starting with 633a094e259945b090fda18510d72f11afe12b0a1f579282734a24c2e455b0ee not found: ID does not exist" containerID="633a094e259945b090fda18510d72f11afe12b0a1f579282734a24c2e455b0ee" Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.886955 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"633a094e259945b090fda18510d72f11afe12b0a1f579282734a24c2e455b0ee"} err="failed to get container status \"633a094e259945b090fda18510d72f11afe12b0a1f579282734a24c2e455b0ee\": rpc error: code = NotFound desc = could not find container \"633a094e259945b090fda18510d72f11afe12b0a1f579282734a24c2e455b0ee\": container with ID starting with 633a094e259945b090fda18510d72f11afe12b0a1f579282734a24c2e455b0ee not found: ID does not exist" Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.886990 4696 scope.go:117] "RemoveContainer" containerID="7c77d3da52a0549309ce8ca707fcc34d55de4d21d9d8040ed15aef36bb950f25" Mar 21 08:47:05 crc kubenswrapper[4696]: E0321 08:47:05.896152 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c77d3da52a0549309ce8ca707fcc34d55de4d21d9d8040ed15aef36bb950f25\": container with ID starting with 7c77d3da52a0549309ce8ca707fcc34d55de4d21d9d8040ed15aef36bb950f25 not found: ID does not exist" containerID="7c77d3da52a0549309ce8ca707fcc34d55de4d21d9d8040ed15aef36bb950f25" Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.896204 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c77d3da52a0549309ce8ca707fcc34d55de4d21d9d8040ed15aef36bb950f25"} err="failed to get container status \"7c77d3da52a0549309ce8ca707fcc34d55de4d21d9d8040ed15aef36bb950f25\": rpc error: code = NotFound desc = could not find container \"7c77d3da52a0549309ce8ca707fcc34d55de4d21d9d8040ed15aef36bb950f25\": container with ID starting with 7c77d3da52a0549309ce8ca707fcc34d55de4d21d9d8040ed15aef36bb950f25 not found: ID does not exist" Mar 21 08:47:05 crc kubenswrapper[4696]: I0321 08:47:05.944729 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.073301 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-ovsdbserver-sb\") pod \"d621abee-2030-44d8-8052-2db726daac53\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.073637 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-dns-svc\") pod \"d621abee-2030-44d8-8052-2db726daac53\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.073708 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-config\") pod \"d621abee-2030-44d8-8052-2db726daac53\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.073810 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52shz\" (UniqueName: \"kubernetes.io/projected/d621abee-2030-44d8-8052-2db726daac53-kube-api-access-52shz\") pod \"d621abee-2030-44d8-8052-2db726daac53\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.073893 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-ovsdbserver-nb\") pod \"d621abee-2030-44d8-8052-2db726daac53\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.073930 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-dns-swift-storage-0\") pod \"d621abee-2030-44d8-8052-2db726daac53\" (UID: \"d621abee-2030-44d8-8052-2db726daac53\") " Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.089802 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d621abee-2030-44d8-8052-2db726daac53-kube-api-access-52shz" (OuterVolumeSpecName: "kube-api-access-52shz") pod "d621abee-2030-44d8-8052-2db726daac53" (UID: "d621abee-2030-44d8-8052-2db726daac53"). InnerVolumeSpecName "kube-api-access-52shz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.168142 4696 generic.go:334] "Generic (PLEG): container finished" podID="64b14a78-860a-48c3-ae69-16cc36c77be8" containerID="3ea87006f13706ba3d8ac0954309bbbc6e34e6fc1350bd3952ec2ae56578a677" exitCode=0 Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.168229 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" event={"ID":"64b14a78-860a-48c3-ae69-16cc36c77be8","Type":"ContainerDied","Data":"3ea87006f13706ba3d8ac0954309bbbc6e34e6fc1350bd3952ec2ae56578a677"} Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.176794 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52shz\" (UniqueName: \"kubernetes.io/projected/d621abee-2030-44d8-8052-2db726daac53-kube-api-access-52shz\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.181063 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"01e35b12-5b6a-4cdd-941f-b2efb0aca029","Type":"ContainerStarted","Data":"687c34dd7f2f4e5d4af31419f662e2ccdee39cc5dc1afd655fe60f77f0207740"} Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.186418 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.187611 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-ggnvq" event={"ID":"d621abee-2030-44d8-8052-2db726daac53","Type":"ContainerDied","Data":"49a26d9ada18a6a9d18f2a4b8fa8f44791c5c82dc9e408f2367f352a3d7c8ff1"} Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.187666 4696 scope.go:117] "RemoveContainer" containerID="78e4f6e2baa3be1cd5da901303d3358ab33a3fad97d133fce46abf24fd584fd0" Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.205660 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d621abee-2030-44d8-8052-2db726daac53" (UID: "d621abee-2030-44d8-8052-2db726daac53"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.232592 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d621abee-2030-44d8-8052-2db726daac53" (UID: "d621abee-2030-44d8-8052-2db726daac53"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.257930 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-config" (OuterVolumeSpecName: "config") pod "d621abee-2030-44d8-8052-2db726daac53" (UID: "d621abee-2030-44d8-8052-2db726daac53"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.279779 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.279825 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.279834 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.282487 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d621abee-2030-44d8-8052-2db726daac53" (UID: "d621abee-2030-44d8-8052-2db726daac53"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.290252 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d621abee-2030-44d8-8052-2db726daac53" (UID: "d621abee-2030-44d8-8052-2db726daac53"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.381981 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.382299 4696 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d621abee-2030-44d8-8052-2db726daac53-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.591420 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ec50e79-4227-4770-b7bc-7dd5f4178fad" path="/var/lib/kubelet/pods/7ec50e79-4227-4770-b7bc-7dd5f4178fad/volumes" Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.594043 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-ggnvq"] Mar 21 08:47:06 crc kubenswrapper[4696]: I0321 08:47:06.594077 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-ggnvq"] Mar 21 08:47:07 crc kubenswrapper[4696]: I0321 08:47:07.228272 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"01e35b12-5b6a-4cdd-941f-b2efb0aca029","Type":"ContainerStarted","Data":"95a5cefb1ffc52985853317004f353f8251f6582d02e1d0f2b7f6ef44f9cc62f"} Mar 21 08:47:07 crc kubenswrapper[4696]: I0321 08:47:07.246995 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2787955f-dbec-430c-91b1-7554a2b20b23","Type":"ContainerStarted","Data":"b9381ccce1e62ee452054520461d74cf2c25b4f9ec56d473065691f0ec43c3f5"} Mar 21 08:47:07 crc kubenswrapper[4696]: I0321 08:47:07.256303 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" event={"ID":"64b14a78-860a-48c3-ae69-16cc36c77be8","Type":"ContainerStarted","Data":"17745124f29da5ac840d31851871596d27883037e3f61a184b8f34cb87fd4a9b"} Mar 21 08:47:07 crc kubenswrapper[4696]: I0321 08:47:07.256488 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:07 crc kubenswrapper[4696]: I0321 08:47:07.288192 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" podStartSLOduration=5.288174717 podStartE2EDuration="5.288174717s" podCreationTimestamp="2026-03-21 08:47:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:47:07.276138501 +0000 UTC m=+1161.397019204" watchObservedRunningTime="2026-03-21 08:47:07.288174717 +0000 UTC m=+1161.409055430" Mar 21 08:47:08 crc kubenswrapper[4696]: I0321 08:47:08.272869 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2787955f-dbec-430c-91b1-7554a2b20b23","Type":"ContainerStarted","Data":"03fad2a910ca1e5e7a105b44ef5ea70d469828c8dd4eb8641230672e8d83c8a9"} Mar 21 08:47:08 crc kubenswrapper[4696]: I0321 08:47:08.273013 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2787955f-dbec-430c-91b1-7554a2b20b23" containerName="glance-log" containerID="cri-o://b9381ccce1e62ee452054520461d74cf2c25b4f9ec56d473065691f0ec43c3f5" gracePeriod=30 Mar 21 08:47:08 crc kubenswrapper[4696]: I0321 08:47:08.273416 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2787955f-dbec-430c-91b1-7554a2b20b23" containerName="glance-httpd" containerID="cri-o://03fad2a910ca1e5e7a105b44ef5ea70d469828c8dd4eb8641230672e8d83c8a9" gracePeriod=30 Mar 21 08:47:08 crc kubenswrapper[4696]: I0321 08:47:08.280753 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="01e35b12-5b6a-4cdd-941f-b2efb0aca029" containerName="glance-log" containerID="cri-o://95a5cefb1ffc52985853317004f353f8251f6582d02e1d0f2b7f6ef44f9cc62f" gracePeriod=30 Mar 21 08:47:08 crc kubenswrapper[4696]: I0321 08:47:08.280950 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"01e35b12-5b6a-4cdd-941f-b2efb0aca029","Type":"ContainerStarted","Data":"4350f916e93b9b8964447a62296ce064804e9883c54cbaf37efdbe6f1c683ccd"} Mar 21 08:47:08 crc kubenswrapper[4696]: I0321 08:47:08.280982 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="01e35b12-5b6a-4cdd-941f-b2efb0aca029" containerName="glance-httpd" containerID="cri-o://4350f916e93b9b8964447a62296ce064804e9883c54cbaf37efdbe6f1c683ccd" gracePeriod=30 Mar 21 08:47:08 crc kubenswrapper[4696]: I0321 08:47:08.307851 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.307810971 podStartE2EDuration="7.307810971s" podCreationTimestamp="2026-03-21 08:47:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:47:08.303010687 +0000 UTC m=+1162.423891410" watchObservedRunningTime="2026-03-21 08:47:08.307810971 +0000 UTC m=+1162.428691684" Mar 21 08:47:08 crc kubenswrapper[4696]: I0321 08:47:08.333318 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.333297342 podStartE2EDuration="6.333297342s" podCreationTimestamp="2026-03-21 08:47:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:47:08.332705785 +0000 UTC m=+1162.453586518" watchObservedRunningTime="2026-03-21 08:47:08.333297342 +0000 UTC m=+1162.454178055" Mar 21 08:47:08 crc kubenswrapper[4696]: I0321 08:47:08.572647 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d621abee-2030-44d8-8052-2db726daac53" path="/var/lib/kubelet/pods/d621abee-2030-44d8-8052-2db726daac53/volumes" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.040016 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.159518 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01e35b12-5b6a-4cdd-941f-b2efb0aca029-logs\") pod \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.159625 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01e35b12-5b6a-4cdd-941f-b2efb0aca029-scripts\") pod \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.159717 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01e35b12-5b6a-4cdd-941f-b2efb0aca029-config-data\") pod \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.159833 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r45cw\" (UniqueName: \"kubernetes.io/projected/01e35b12-5b6a-4cdd-941f-b2efb0aca029-kube-api-access-r45cw\") pod \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.159867 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/01e35b12-5b6a-4cdd-941f-b2efb0aca029-httpd-run\") pod \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.159995 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") pod \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.160034 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01e35b12-5b6a-4cdd-941f-b2efb0aca029-combined-ca-bundle\") pod \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\" (UID: \"01e35b12-5b6a-4cdd-941f-b2efb0aca029\") " Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.170690 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01e35b12-5b6a-4cdd-941f-b2efb0aca029-logs" (OuterVolumeSpecName: "logs") pod "01e35b12-5b6a-4cdd-941f-b2efb0aca029" (UID: "01e35b12-5b6a-4cdd-941f-b2efb0aca029"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.171021 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01e35b12-5b6a-4cdd-941f-b2efb0aca029-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "01e35b12-5b6a-4cdd-941f-b2efb0aca029" (UID: "01e35b12-5b6a-4cdd-941f-b2efb0aca029"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.179256 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01e35b12-5b6a-4cdd-941f-b2efb0aca029-scripts" (OuterVolumeSpecName: "scripts") pod "01e35b12-5b6a-4cdd-941f-b2efb0aca029" (UID: "01e35b12-5b6a-4cdd-941f-b2efb0aca029"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.179378 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01e35b12-5b6a-4cdd-941f-b2efb0aca029-kube-api-access-r45cw" (OuterVolumeSpecName: "kube-api-access-r45cw") pod "01e35b12-5b6a-4cdd-941f-b2efb0aca029" (UID: "01e35b12-5b6a-4cdd-941f-b2efb0aca029"). InnerVolumeSpecName "kube-api-access-r45cw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.187644 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205" (OuterVolumeSpecName: "glance") pod "01e35b12-5b6a-4cdd-941f-b2efb0aca029" (UID: "01e35b12-5b6a-4cdd-941f-b2efb0aca029"). InnerVolumeSpecName "pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.214861 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01e35b12-5b6a-4cdd-941f-b2efb0aca029-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "01e35b12-5b6a-4cdd-941f-b2efb0aca029" (UID: "01e35b12-5b6a-4cdd-941f-b2efb0aca029"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.262298 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01e35b12-5b6a-4cdd-941f-b2efb0aca029-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.262331 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r45cw\" (UniqueName: \"kubernetes.io/projected/01e35b12-5b6a-4cdd-941f-b2efb0aca029-kube-api-access-r45cw\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.262357 4696 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") on node \"crc\" " Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.262367 4696 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/01e35b12-5b6a-4cdd-941f-b2efb0aca029-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.262376 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01e35b12-5b6a-4cdd-941f-b2efb0aca029-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.262389 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01e35b12-5b6a-4cdd-941f-b2efb0aca029-logs\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.267943 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01e35b12-5b6a-4cdd-941f-b2efb0aca029-config-data" (OuterVolumeSpecName: "config-data") pod "01e35b12-5b6a-4cdd-941f-b2efb0aca029" (UID: "01e35b12-5b6a-4cdd-941f-b2efb0aca029"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.305672 4696 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.305840 4696 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205") on node "crc" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.325060 4696 generic.go:334] "Generic (PLEG): container finished" podID="01e35b12-5b6a-4cdd-941f-b2efb0aca029" containerID="4350f916e93b9b8964447a62296ce064804e9883c54cbaf37efdbe6f1c683ccd" exitCode=0 Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.325087 4696 generic.go:334] "Generic (PLEG): container finished" podID="01e35b12-5b6a-4cdd-941f-b2efb0aca029" containerID="95a5cefb1ffc52985853317004f353f8251f6582d02e1d0f2b7f6ef44f9cc62f" exitCode=143 Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.325127 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"01e35b12-5b6a-4cdd-941f-b2efb0aca029","Type":"ContainerDied","Data":"4350f916e93b9b8964447a62296ce064804e9883c54cbaf37efdbe6f1c683ccd"} Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.325152 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"01e35b12-5b6a-4cdd-941f-b2efb0aca029","Type":"ContainerDied","Data":"95a5cefb1ffc52985853317004f353f8251f6582d02e1d0f2b7f6ef44f9cc62f"} Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.325163 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"01e35b12-5b6a-4cdd-941f-b2efb0aca029","Type":"ContainerDied","Data":"687c34dd7f2f4e5d4af31419f662e2ccdee39cc5dc1afd655fe60f77f0207740"} Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.325180 4696 scope.go:117] "RemoveContainer" containerID="4350f916e93b9b8964447a62296ce064804e9883c54cbaf37efdbe6f1c683ccd" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.325291 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.348692 4696 generic.go:334] "Generic (PLEG): container finished" podID="50c2c69e-b049-4b2a-abca-6943a24bdfee" containerID="027b0a845d9b25edaa4aec6a90d1a489fea68982c21b4e5196545686f124576a" exitCode=0 Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.348780 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2pgpc" event={"ID":"50c2c69e-b049-4b2a-abca-6943a24bdfee","Type":"ContainerDied","Data":"027b0a845d9b25edaa4aec6a90d1a489fea68982c21b4e5196545686f124576a"} Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.352057 4696 generic.go:334] "Generic (PLEG): container finished" podID="2787955f-dbec-430c-91b1-7554a2b20b23" containerID="03fad2a910ca1e5e7a105b44ef5ea70d469828c8dd4eb8641230672e8d83c8a9" exitCode=0 Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.352074 4696 generic.go:334] "Generic (PLEG): container finished" podID="2787955f-dbec-430c-91b1-7554a2b20b23" containerID="b9381ccce1e62ee452054520461d74cf2c25b4f9ec56d473065691f0ec43c3f5" exitCode=143 Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.352090 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2787955f-dbec-430c-91b1-7554a2b20b23","Type":"ContainerDied","Data":"03fad2a910ca1e5e7a105b44ef5ea70d469828c8dd4eb8641230672e8d83c8a9"} Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.352104 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2787955f-dbec-430c-91b1-7554a2b20b23","Type":"ContainerDied","Data":"b9381ccce1e62ee452054520461d74cf2c25b4f9ec56d473065691f0ec43c3f5"} Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.393723 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01e35b12-5b6a-4cdd-941f-b2efb0aca029-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.393766 4696 reconciler_common.go:293] "Volume detached for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.400896 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.426362 4696 scope.go:117] "RemoveContainer" containerID="95a5cefb1ffc52985853317004f353f8251f6582d02e1d0f2b7f6ef44f9cc62f" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.442229 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.488178 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 08:47:09 crc kubenswrapper[4696]: E0321 08:47:09.492910 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d621abee-2030-44d8-8052-2db726daac53" containerName="init" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.492965 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d621abee-2030-44d8-8052-2db726daac53" containerName="init" Mar 21 08:47:09 crc kubenswrapper[4696]: E0321 08:47:09.492992 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01e35b12-5b6a-4cdd-941f-b2efb0aca029" containerName="glance-log" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.493001 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="01e35b12-5b6a-4cdd-941f-b2efb0aca029" containerName="glance-log" Mar 21 08:47:09 crc kubenswrapper[4696]: E0321 08:47:09.493049 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01e35b12-5b6a-4cdd-941f-b2efb0aca029" containerName="glance-httpd" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.493060 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="01e35b12-5b6a-4cdd-941f-b2efb0aca029" containerName="glance-httpd" Mar 21 08:47:09 crc kubenswrapper[4696]: E0321 08:47:09.493078 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ec50e79-4227-4770-b7bc-7dd5f4178fad" containerName="init" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.493086 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ec50e79-4227-4770-b7bc-7dd5f4178fad" containerName="init" Mar 21 08:47:09 crc kubenswrapper[4696]: E0321 08:47:09.493117 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ec50e79-4227-4770-b7bc-7dd5f4178fad" containerName="dnsmasq-dns" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.493125 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ec50e79-4227-4770-b7bc-7dd5f4178fad" containerName="dnsmasq-dns" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.493550 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="d621abee-2030-44d8-8052-2db726daac53" containerName="init" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.493590 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="01e35b12-5b6a-4cdd-941f-b2efb0aca029" containerName="glance-log" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.493608 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="01e35b12-5b6a-4cdd-941f-b2efb0aca029" containerName="glance-httpd" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.493631 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ec50e79-4227-4770-b7bc-7dd5f4178fad" containerName="dnsmasq-dns" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.495055 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.496883 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.501360 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.539391 4696 scope.go:117] "RemoveContainer" containerID="4350f916e93b9b8964447a62296ce064804e9883c54cbaf37efdbe6f1c683ccd" Mar 21 08:47:09 crc kubenswrapper[4696]: E0321 08:47:09.540747 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4350f916e93b9b8964447a62296ce064804e9883c54cbaf37efdbe6f1c683ccd\": container with ID starting with 4350f916e93b9b8964447a62296ce064804e9883c54cbaf37efdbe6f1c683ccd not found: ID does not exist" containerID="4350f916e93b9b8964447a62296ce064804e9883c54cbaf37efdbe6f1c683ccd" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.540778 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4350f916e93b9b8964447a62296ce064804e9883c54cbaf37efdbe6f1c683ccd"} err="failed to get container status \"4350f916e93b9b8964447a62296ce064804e9883c54cbaf37efdbe6f1c683ccd\": rpc error: code = NotFound desc = could not find container \"4350f916e93b9b8964447a62296ce064804e9883c54cbaf37efdbe6f1c683ccd\": container with ID starting with 4350f916e93b9b8964447a62296ce064804e9883c54cbaf37efdbe6f1c683ccd not found: ID does not exist" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.540798 4696 scope.go:117] "RemoveContainer" containerID="95a5cefb1ffc52985853317004f353f8251f6582d02e1d0f2b7f6ef44f9cc62f" Mar 21 08:47:09 crc kubenswrapper[4696]: E0321 08:47:09.544973 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95a5cefb1ffc52985853317004f353f8251f6582d02e1d0f2b7f6ef44f9cc62f\": container with ID starting with 95a5cefb1ffc52985853317004f353f8251f6582d02e1d0f2b7f6ef44f9cc62f not found: ID does not exist" containerID="95a5cefb1ffc52985853317004f353f8251f6582d02e1d0f2b7f6ef44f9cc62f" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.545026 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95a5cefb1ffc52985853317004f353f8251f6582d02e1d0f2b7f6ef44f9cc62f"} err="failed to get container status \"95a5cefb1ffc52985853317004f353f8251f6582d02e1d0f2b7f6ef44f9cc62f\": rpc error: code = NotFound desc = could not find container \"95a5cefb1ffc52985853317004f353f8251f6582d02e1d0f2b7f6ef44f9cc62f\": container with ID starting with 95a5cefb1ffc52985853317004f353f8251f6582d02e1d0f2b7f6ef44f9cc62f not found: ID does not exist" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.545059 4696 scope.go:117] "RemoveContainer" containerID="4350f916e93b9b8964447a62296ce064804e9883c54cbaf37efdbe6f1c683ccd" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.545525 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4350f916e93b9b8964447a62296ce064804e9883c54cbaf37efdbe6f1c683ccd"} err="failed to get container status \"4350f916e93b9b8964447a62296ce064804e9883c54cbaf37efdbe6f1c683ccd\": rpc error: code = NotFound desc = could not find container \"4350f916e93b9b8964447a62296ce064804e9883c54cbaf37efdbe6f1c683ccd\": container with ID starting with 4350f916e93b9b8964447a62296ce064804e9883c54cbaf37efdbe6f1c683ccd not found: ID does not exist" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.545554 4696 scope.go:117] "RemoveContainer" containerID="95a5cefb1ffc52985853317004f353f8251f6582d02e1d0f2b7f6ef44f9cc62f" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.547987 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95a5cefb1ffc52985853317004f353f8251f6582d02e1d0f2b7f6ef44f9cc62f"} err="failed to get container status \"95a5cefb1ffc52985853317004f353f8251f6582d02e1d0f2b7f6ef44f9cc62f\": rpc error: code = NotFound desc = could not find container \"95a5cefb1ffc52985853317004f353f8251f6582d02e1d0f2b7f6ef44f9cc62f\": container with ID starting with 95a5cefb1ffc52985853317004f353f8251f6582d02e1d0f2b7f6ef44f9cc62f not found: ID does not exist" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.599978 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/296e9317-a507-4e57-8ee3-d42a091f866c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.600078 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/296e9317-a507-4e57-8ee3-d42a091f866c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.600132 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.600201 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/296e9317-a507-4e57-8ee3-d42a091f866c-logs\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.600247 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/296e9317-a507-4e57-8ee3-d42a091f866c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.600290 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzr66\" (UniqueName: \"kubernetes.io/projected/296e9317-a507-4e57-8ee3-d42a091f866c-kube-api-access-fzr66\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.600326 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/296e9317-a507-4e57-8ee3-d42a091f866c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.635078 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.702253 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/296e9317-a507-4e57-8ee3-d42a091f866c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.702350 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/296e9317-a507-4e57-8ee3-d42a091f866c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.702388 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.702449 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/296e9317-a507-4e57-8ee3-d42a091f866c-logs\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.702484 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/296e9317-a507-4e57-8ee3-d42a091f866c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.702547 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzr66\" (UniqueName: \"kubernetes.io/projected/296e9317-a507-4e57-8ee3-d42a091f866c-kube-api-access-fzr66\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.703421 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/296e9317-a507-4e57-8ee3-d42a091f866c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.704612 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/296e9317-a507-4e57-8ee3-d42a091f866c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.706265 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/296e9317-a507-4e57-8ee3-d42a091f866c-logs\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.710430 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/296e9317-a507-4e57-8ee3-d42a091f866c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.711318 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/296e9317-a507-4e57-8ee3-d42a091f866c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.717511 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.717542 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/36feb21b6af5aad5e90205d24e3f1d239a0e9aaaadf2ca8cf7f1809b63388e1a/globalmount\"" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.719597 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/296e9317-a507-4e57-8ee3-d42a091f866c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.720576 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzr66\" (UniqueName: \"kubernetes.io/projected/296e9317-a507-4e57-8ee3-d42a091f866c-kube-api-access-fzr66\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.765015 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") pod \"glance-default-internal-api-0\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.804466 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2787955f-dbec-430c-91b1-7554a2b20b23-httpd-run\") pod \"2787955f-dbec-430c-91b1-7554a2b20b23\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.804572 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2787955f-dbec-430c-91b1-7554a2b20b23-config-data\") pod \"2787955f-dbec-430c-91b1-7554a2b20b23\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.804615 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2787955f-dbec-430c-91b1-7554a2b20b23-combined-ca-bundle\") pod \"2787955f-dbec-430c-91b1-7554a2b20b23\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.804729 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fzxp\" (UniqueName: \"kubernetes.io/projected/2787955f-dbec-430c-91b1-7554a2b20b23-kube-api-access-2fzxp\") pod \"2787955f-dbec-430c-91b1-7554a2b20b23\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.804768 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2787955f-dbec-430c-91b1-7554a2b20b23-scripts\") pod \"2787955f-dbec-430c-91b1-7554a2b20b23\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.804909 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") pod \"2787955f-dbec-430c-91b1-7554a2b20b23\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.804932 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2787955f-dbec-430c-91b1-7554a2b20b23-logs\") pod \"2787955f-dbec-430c-91b1-7554a2b20b23\" (UID: \"2787955f-dbec-430c-91b1-7554a2b20b23\") " Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.805654 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2787955f-dbec-430c-91b1-7554a2b20b23-logs" (OuterVolumeSpecName: "logs") pod "2787955f-dbec-430c-91b1-7554a2b20b23" (UID: "2787955f-dbec-430c-91b1-7554a2b20b23"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.805784 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2787955f-dbec-430c-91b1-7554a2b20b23-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2787955f-dbec-430c-91b1-7554a2b20b23" (UID: "2787955f-dbec-430c-91b1-7554a2b20b23"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.812974 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2787955f-dbec-430c-91b1-7554a2b20b23-kube-api-access-2fzxp" (OuterVolumeSpecName: "kube-api-access-2fzxp") pod "2787955f-dbec-430c-91b1-7554a2b20b23" (UID: "2787955f-dbec-430c-91b1-7554a2b20b23"). InnerVolumeSpecName "kube-api-access-2fzxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.827009 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2787955f-dbec-430c-91b1-7554a2b20b23-scripts" (OuterVolumeSpecName: "scripts") pod "2787955f-dbec-430c-91b1-7554a2b20b23" (UID: "2787955f-dbec-430c-91b1-7554a2b20b23"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.859011 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2787955f-dbec-430c-91b1-7554a2b20b23-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2787955f-dbec-430c-91b1-7554a2b20b23" (UID: "2787955f-dbec-430c-91b1-7554a2b20b23"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.869106 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2787955f-dbec-430c-91b1-7554a2b20b23-config-data" (OuterVolumeSpecName: "config-data") pod "2787955f-dbec-430c-91b1-7554a2b20b23" (UID: "2787955f-dbec-430c-91b1-7554a2b20b23"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.913064 4696 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2787955f-dbec-430c-91b1-7554a2b20b23-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.913095 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2787955f-dbec-430c-91b1-7554a2b20b23-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.913104 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2787955f-dbec-430c-91b1-7554a2b20b23-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.913114 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fzxp\" (UniqueName: \"kubernetes.io/projected/2787955f-dbec-430c-91b1-7554a2b20b23-kube-api-access-2fzxp\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.913122 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2787955f-dbec-430c-91b1-7554a2b20b23-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.913131 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2787955f-dbec-430c-91b1-7554a2b20b23-logs\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:09 crc kubenswrapper[4696]: I0321 08:47:09.923498 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.081656 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689" (OuterVolumeSpecName: "glance") pod "2787955f-dbec-430c-91b1-7554a2b20b23" (UID: "2787955f-dbec-430c-91b1-7554a2b20b23"). InnerVolumeSpecName "pvc-984a8dcb-40a1-4f39-bf33-0005b91df689". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.116382 4696 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") on node \"crc\" " Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.201905 4696 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.202397 4696 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-984a8dcb-40a1-4f39-bf33-0005b91df689" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689") on node "crc" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.218448 4696 reconciler_common.go:293] "Volume detached for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.367158 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.367200 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2787955f-dbec-430c-91b1-7554a2b20b23","Type":"ContainerDied","Data":"bc0f472e71fa433353954b6d8102751a93e769117f3aa7371216a6fc2fc8ed26"} Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.367233 4696 scope.go:117] "RemoveContainer" containerID="03fad2a910ca1e5e7a105b44ef5ea70d469828c8dd4eb8641230672e8d83c8a9" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.425643 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.451998 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.467583 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 08:47:10 crc kubenswrapper[4696]: E0321 08:47:10.468047 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2787955f-dbec-430c-91b1-7554a2b20b23" containerName="glance-log" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.468062 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2787955f-dbec-430c-91b1-7554a2b20b23" containerName="glance-log" Mar 21 08:47:10 crc kubenswrapper[4696]: E0321 08:47:10.468089 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2787955f-dbec-430c-91b1-7554a2b20b23" containerName="glance-httpd" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.468095 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2787955f-dbec-430c-91b1-7554a2b20b23" containerName="glance-httpd" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.468308 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="2787955f-dbec-430c-91b1-7554a2b20b23" containerName="glance-httpd" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.468331 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="2787955f-dbec-430c-91b1-7554a2b20b23" containerName="glance-log" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.469400 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.473230 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.490592 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.551685 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01e35b12-5b6a-4cdd-941f-b2efb0aca029" path="/var/lib/kubelet/pods/01e35b12-5b6a-4cdd-941f-b2efb0aca029/volumes" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.552530 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2787955f-dbec-430c-91b1-7554a2b20b23" path="/var/lib/kubelet/pods/2787955f-dbec-430c-91b1-7554a2b20b23/volumes" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.591937 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 08:47:10 crc kubenswrapper[4696]: E0321 08:47:10.592606 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data glance httpd-run kube-api-access-8t8pk logs scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-external-api-0" podUID="35e9408b-dab7-4b52-befd-c6e93380ef6b" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.635549 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35e9408b-dab7-4b52-befd-c6e93380ef6b-logs\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.635600 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t8pk\" (UniqueName: \"kubernetes.io/projected/35e9408b-dab7-4b52-befd-c6e93380ef6b-kube-api-access-8t8pk\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.635697 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e9408b-dab7-4b52-befd-c6e93380ef6b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.635719 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35e9408b-dab7-4b52-befd-c6e93380ef6b-scripts\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.635743 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.635762 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35e9408b-dab7-4b52-befd-c6e93380ef6b-config-data\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.635776 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/35e9408b-dab7-4b52-befd-c6e93380ef6b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.682067 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.730786 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.738557 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t8pk\" (UniqueName: \"kubernetes.io/projected/35e9408b-dab7-4b52-befd-c6e93380ef6b-kube-api-access-8t8pk\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.741988 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e9408b-dab7-4b52-befd-c6e93380ef6b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.742053 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35e9408b-dab7-4b52-befd-c6e93380ef6b-scripts\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.742100 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.742139 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35e9408b-dab7-4b52-befd-c6e93380ef6b-config-data\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.742168 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/35e9408b-dab7-4b52-befd-c6e93380ef6b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.742353 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35e9408b-dab7-4b52-befd-c6e93380ef6b-logs\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.742900 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35e9408b-dab7-4b52-befd-c6e93380ef6b-logs\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.743202 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/35e9408b-dab7-4b52-befd-c6e93380ef6b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.745229 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.745251 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0bb60238debc80c6ed39528f98bb598b47b2c7acc3e22f58b887f712b724eea4/globalmount\"" pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.747976 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35e9408b-dab7-4b52-befd-c6e93380ef6b-config-data\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.749017 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e9408b-dab7-4b52-befd-c6e93380ef6b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.751005 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35e9408b-dab7-4b52-befd-c6e93380ef6b-scripts\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.753856 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t8pk\" (UniqueName: \"kubernetes.io/projected/35e9408b-dab7-4b52-befd-c6e93380ef6b-kube-api-access-8t8pk\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:10 crc kubenswrapper[4696]: I0321 08:47:10.799990 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") pod \"glance-default-external-api-0\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.378291 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.401983 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.562527 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35e9408b-dab7-4b52-befd-c6e93380ef6b-logs\") pod \"35e9408b-dab7-4b52-befd-c6e93380ef6b\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.562675 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35e9408b-dab7-4b52-befd-c6e93380ef6b-config-data\") pod \"35e9408b-dab7-4b52-befd-c6e93380ef6b\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.562807 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") pod \"35e9408b-dab7-4b52-befd-c6e93380ef6b\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.562866 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/35e9408b-dab7-4b52-befd-c6e93380ef6b-httpd-run\") pod \"35e9408b-dab7-4b52-befd-c6e93380ef6b\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.562882 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35e9408b-dab7-4b52-befd-c6e93380ef6b-logs" (OuterVolumeSpecName: "logs") pod "35e9408b-dab7-4b52-befd-c6e93380ef6b" (UID: "35e9408b-dab7-4b52-befd-c6e93380ef6b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.562929 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8t8pk\" (UniqueName: \"kubernetes.io/projected/35e9408b-dab7-4b52-befd-c6e93380ef6b-kube-api-access-8t8pk\") pod \"35e9408b-dab7-4b52-befd-c6e93380ef6b\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.563153 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35e9408b-dab7-4b52-befd-c6e93380ef6b-scripts\") pod \"35e9408b-dab7-4b52-befd-c6e93380ef6b\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.563185 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e9408b-dab7-4b52-befd-c6e93380ef6b-combined-ca-bundle\") pod \"35e9408b-dab7-4b52-befd-c6e93380ef6b\" (UID: \"35e9408b-dab7-4b52-befd-c6e93380ef6b\") " Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.563240 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35e9408b-dab7-4b52-befd-c6e93380ef6b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "35e9408b-dab7-4b52-befd-c6e93380ef6b" (UID: "35e9408b-dab7-4b52-befd-c6e93380ef6b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.564361 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35e9408b-dab7-4b52-befd-c6e93380ef6b-logs\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.564384 4696 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/35e9408b-dab7-4b52-befd-c6e93380ef6b-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.572122 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e9408b-dab7-4b52-befd-c6e93380ef6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35e9408b-dab7-4b52-befd-c6e93380ef6b" (UID: "35e9408b-dab7-4b52-befd-c6e93380ef6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.572182 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e9408b-dab7-4b52-befd-c6e93380ef6b-scripts" (OuterVolumeSpecName: "scripts") pod "35e9408b-dab7-4b52-befd-c6e93380ef6b" (UID: "35e9408b-dab7-4b52-befd-c6e93380ef6b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.572212 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35e9408b-dab7-4b52-befd-c6e93380ef6b-kube-api-access-8t8pk" (OuterVolumeSpecName: "kube-api-access-8t8pk") pod "35e9408b-dab7-4b52-befd-c6e93380ef6b" (UID: "35e9408b-dab7-4b52-befd-c6e93380ef6b"). InnerVolumeSpecName "kube-api-access-8t8pk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.572336 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e9408b-dab7-4b52-befd-c6e93380ef6b-config-data" (OuterVolumeSpecName: "config-data") pod "35e9408b-dab7-4b52-befd-c6e93380ef6b" (UID: "35e9408b-dab7-4b52-befd-c6e93380ef6b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.583515 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689" (OuterVolumeSpecName: "glance") pod "35e9408b-dab7-4b52-befd-c6e93380ef6b" (UID: "35e9408b-dab7-4b52-befd-c6e93380ef6b"). InnerVolumeSpecName "pvc-984a8dcb-40a1-4f39-bf33-0005b91df689". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.666835 4696 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") on node \"crc\" " Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.666873 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8t8pk\" (UniqueName: \"kubernetes.io/projected/35e9408b-dab7-4b52-befd-c6e93380ef6b-kube-api-access-8t8pk\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.666889 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35e9408b-dab7-4b52-befd-c6e93380ef6b-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.666901 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e9408b-dab7-4b52-befd-c6e93380ef6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.666912 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35e9408b-dab7-4b52-befd-c6e93380ef6b-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.697960 4696 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.698303 4696 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-984a8dcb-40a1-4f39-bf33-0005b91df689" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689") on node "crc" Mar 21 08:47:11 crc kubenswrapper[4696]: I0321 08:47:11.770148 4696 reconciler_common.go:293] "Volume detached for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.395591 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.445184 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.452665 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.478001 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.481705 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.484243 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.484885 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.507693 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.547311 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35e9408b-dab7-4b52-befd-c6e93380ef6b" path="/var/lib/kubelet/pods/35e9408b-dab7-4b52-befd-c6e93380ef6b/volumes" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.589746 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.589863 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-scripts\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.589904 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xpbn\" (UniqueName: \"kubernetes.io/projected/02fca1e0-3685-4d5b-86f9-bef0511876df-kube-api-access-6xpbn\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.589921 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.589940 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/02fca1e0-3685-4d5b-86f9-bef0511876df-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.590184 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-config-data\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.590295 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.590718 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02fca1e0-3685-4d5b-86f9-bef0511876df-logs\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.692148 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-config-data\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.692236 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.692376 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02fca1e0-3685-4d5b-86f9-bef0511876df-logs\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.692498 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.692580 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-scripts\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.692643 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xpbn\" (UniqueName: \"kubernetes.io/projected/02fca1e0-3685-4d5b-86f9-bef0511876df-kube-api-access-6xpbn\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.692682 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.692700 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/02fca1e0-3685-4d5b-86f9-bef0511876df-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.695196 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02fca1e0-3685-4d5b-86f9-bef0511876df-logs\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.697210 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-scripts\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.698931 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-config-data\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.699161 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/02fca1e0-3685-4d5b-86f9-bef0511876df-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.699183 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.700413 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.700438 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0bb60238debc80c6ed39528f98bb598b47b2c7acc3e22f58b887f712b724eea4/globalmount\"" pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.704440 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.720252 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xpbn\" (UniqueName: \"kubernetes.io/projected/02fca1e0-3685-4d5b-86f9-bef0511876df-kube-api-access-6xpbn\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.740631 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") pod \"glance-default-external-api-0\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " pod="openstack/glance-default-external-api-0" Mar 21 08:47:12 crc kubenswrapper[4696]: I0321 08:47:12.799949 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 08:47:13 crc kubenswrapper[4696]: I0321 08:47:13.043993 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:13 crc kubenswrapper[4696]: I0321 08:47:13.132988 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-48vx9"] Mar 21 08:47:13 crc kubenswrapper[4696]: I0321 08:47:13.133213 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-48vx9" podUID="782d4c3b-a1bf-4e02-ba5b-0274f885c26b" containerName="dnsmasq-dns" containerID="cri-o://7e7507b773e14ca3d089aa8075b93fec3216569eeeb079b66f4723631ab2cfc1" gracePeriod=10 Mar 21 08:47:13 crc kubenswrapper[4696]: I0321 08:47:13.438354 4696 generic.go:334] "Generic (PLEG): container finished" podID="782d4c3b-a1bf-4e02-ba5b-0274f885c26b" containerID="7e7507b773e14ca3d089aa8075b93fec3216569eeeb079b66f4723631ab2cfc1" exitCode=0 Mar 21 08:47:13 crc kubenswrapper[4696]: I0321 08:47:13.438409 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-48vx9" event={"ID":"782d4c3b-a1bf-4e02-ba5b-0274f885c26b","Type":"ContainerDied","Data":"7e7507b773e14ca3d089aa8075b93fec3216569eeeb079b66f4723631ab2cfc1"} Mar 21 08:47:16 crc kubenswrapper[4696]: I0321 08:47:16.581571 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-48vx9" podUID="782d4c3b-a1bf-4e02-ba5b-0274f885c26b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.134:5353: connect: connection refused" Mar 21 08:47:17 crc kubenswrapper[4696]: W0321 08:47:17.084318 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod296e9317_a507_4e57_8ee3_d42a091f866c.slice/crio-31dd2530f6846337a05986d4f3310cf13376e7e01a1160b15069c17b3fcb081a WatchSource:0}: Error finding container 31dd2530f6846337a05986d4f3310cf13376e7e01a1160b15069c17b3fcb081a: Status 404 returned error can't find the container with id 31dd2530f6846337a05986d4f3310cf13376e7e01a1160b15069c17b3fcb081a Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.218993 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.233324 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-credential-keys\") pod \"50c2c69e-b049-4b2a-abca-6943a24bdfee\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.233581 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-scripts\") pod \"50c2c69e-b049-4b2a-abca-6943a24bdfee\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.233612 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-config-data\") pod \"50c2c69e-b049-4b2a-abca-6943a24bdfee\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.233692 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-combined-ca-bundle\") pod \"50c2c69e-b049-4b2a-abca-6943a24bdfee\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.233748 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-fernet-keys\") pod \"50c2c69e-b049-4b2a-abca-6943a24bdfee\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.233943 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t92jr\" (UniqueName: \"kubernetes.io/projected/50c2c69e-b049-4b2a-abca-6943a24bdfee-kube-api-access-t92jr\") pod \"50c2c69e-b049-4b2a-abca-6943a24bdfee\" (UID: \"50c2c69e-b049-4b2a-abca-6943a24bdfee\") " Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.253240 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-scripts" (OuterVolumeSpecName: "scripts") pod "50c2c69e-b049-4b2a-abca-6943a24bdfee" (UID: "50c2c69e-b049-4b2a-abca-6943a24bdfee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.255206 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "50c2c69e-b049-4b2a-abca-6943a24bdfee" (UID: "50c2c69e-b049-4b2a-abca-6943a24bdfee"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.262021 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50c2c69e-b049-4b2a-abca-6943a24bdfee-kube-api-access-t92jr" (OuterVolumeSpecName: "kube-api-access-t92jr") pod "50c2c69e-b049-4b2a-abca-6943a24bdfee" (UID: "50c2c69e-b049-4b2a-abca-6943a24bdfee"). InnerVolumeSpecName "kube-api-access-t92jr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.262337 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "50c2c69e-b049-4b2a-abca-6943a24bdfee" (UID: "50c2c69e-b049-4b2a-abca-6943a24bdfee"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.290965 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50c2c69e-b049-4b2a-abca-6943a24bdfee" (UID: "50c2c69e-b049-4b2a-abca-6943a24bdfee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.292619 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-config-data" (OuterVolumeSpecName: "config-data") pod "50c2c69e-b049-4b2a-abca-6943a24bdfee" (UID: "50c2c69e-b049-4b2a-abca-6943a24bdfee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.336943 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t92jr\" (UniqueName: \"kubernetes.io/projected/50c2c69e-b049-4b2a-abca-6943a24bdfee-kube-api-access-t92jr\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.336979 4696 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.336989 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.336997 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.337008 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.337015 4696 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/50c2c69e-b049-4b2a-abca-6943a24bdfee-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.480801 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2pgpc" event={"ID":"50c2c69e-b049-4b2a-abca-6943a24bdfee","Type":"ContainerDied","Data":"8c1e7a3aba219c87466ffc5f362517623ce0a2ce62f094e2d24dc64a00d53b5f"} Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.480844 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2pgpc" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.481194 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c1e7a3aba219c87466ffc5f362517623ce0a2ce62f094e2d24dc64a00d53b5f" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.482787 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"296e9317-a507-4e57-8ee3-d42a091f866c","Type":"ContainerStarted","Data":"31dd2530f6846337a05986d4f3310cf13376e7e01a1160b15069c17b3fcb081a"} Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.712028 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-48vx9" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.715698 4696 scope.go:117] "RemoveContainer" containerID="b9381ccce1e62ee452054520461d74cf2c25b4f9ec56d473065691f0ec43c3f5" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.743164 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-ovsdbserver-sb\") pod \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\" (UID: \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\") " Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.743636 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-ovsdbserver-nb\") pod \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\" (UID: \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\") " Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.743733 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-config\") pod \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\" (UID: \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\") " Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.743790 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-dns-svc\") pod \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\" (UID: \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\") " Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.743905 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzkcc\" (UniqueName: \"kubernetes.io/projected/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-kube-api-access-zzkcc\") pod \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\" (UID: \"782d4c3b-a1bf-4e02-ba5b-0274f885c26b\") " Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.753375 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-kube-api-access-zzkcc" (OuterVolumeSpecName: "kube-api-access-zzkcc") pod "782d4c3b-a1bf-4e02-ba5b-0274f885c26b" (UID: "782d4c3b-a1bf-4e02-ba5b-0274f885c26b"). InnerVolumeSpecName "kube-api-access-zzkcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.784414 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "782d4c3b-a1bf-4e02-ba5b-0274f885c26b" (UID: "782d4c3b-a1bf-4e02-ba5b-0274f885c26b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.799622 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "782d4c3b-a1bf-4e02-ba5b-0274f885c26b" (UID: "782d4c3b-a1bf-4e02-ba5b-0274f885c26b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.810591 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "782d4c3b-a1bf-4e02-ba5b-0274f885c26b" (UID: "782d4c3b-a1bf-4e02-ba5b-0274f885c26b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.824829 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-config" (OuterVolumeSpecName: "config") pod "782d4c3b-a1bf-4e02-ba5b-0274f885c26b" (UID: "782d4c3b-a1bf-4e02-ba5b-0274f885c26b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.845454 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.845490 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.845500 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzkcc\" (UniqueName: \"kubernetes.io/projected/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-kube-api-access-zzkcc\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.845511 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:17 crc kubenswrapper[4696]: I0321 08:47:17.845520 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/782d4c3b-a1bf-4e02-ba5b-0274f885c26b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.287093 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-2pgpc"] Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.295504 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-2pgpc"] Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.389649 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-cwvf9"] Mar 21 08:47:18 crc kubenswrapper[4696]: E0321 08:47:18.390179 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="782d4c3b-a1bf-4e02-ba5b-0274f885c26b" containerName="init" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.390196 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="782d4c3b-a1bf-4e02-ba5b-0274f885c26b" containerName="init" Mar 21 08:47:18 crc kubenswrapper[4696]: E0321 08:47:18.390214 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50c2c69e-b049-4b2a-abca-6943a24bdfee" containerName="keystone-bootstrap" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.390221 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="50c2c69e-b049-4b2a-abca-6943a24bdfee" containerName="keystone-bootstrap" Mar 21 08:47:18 crc kubenswrapper[4696]: E0321 08:47:18.390234 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="782d4c3b-a1bf-4e02-ba5b-0274f885c26b" containerName="dnsmasq-dns" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.390241 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="782d4c3b-a1bf-4e02-ba5b-0274f885c26b" containerName="dnsmasq-dns" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.390463 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="782d4c3b-a1bf-4e02-ba5b-0274f885c26b" containerName="dnsmasq-dns" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.390483 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="50c2c69e-b049-4b2a-abca-6943a24bdfee" containerName="keystone-bootstrap" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.391352 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.393796 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.393951 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.394157 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.394179 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-s9frx" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.397949 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-cwvf9"] Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.459482 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-fernet-keys\") pod \"keystone-bootstrap-cwvf9\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.459552 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-credential-keys\") pod \"keystone-bootstrap-cwvf9\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.459624 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-combined-ca-bundle\") pod \"keystone-bootstrap-cwvf9\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.459669 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpnx6\" (UniqueName: \"kubernetes.io/projected/b8556431-841c-4372-8c93-fc3649d9b0f0-kube-api-access-wpnx6\") pod \"keystone-bootstrap-cwvf9\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.459728 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-scripts\") pod \"keystone-bootstrap-cwvf9\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.459757 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-config-data\") pod \"keystone-bootstrap-cwvf9\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.501158 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-48vx9" event={"ID":"782d4c3b-a1bf-4e02-ba5b-0274f885c26b","Type":"ContainerDied","Data":"cb2812fb454ed3253abf294282aef582ecbccda536b8f0dcfb017acb0015c06d"} Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.501209 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-48vx9" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.549971 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50c2c69e-b049-4b2a-abca-6943a24bdfee" path="/var/lib/kubelet/pods/50c2c69e-b049-4b2a-abca-6943a24bdfee/volumes" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.550588 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-48vx9"] Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.551191 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-48vx9"] Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.560997 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-fernet-keys\") pod \"keystone-bootstrap-cwvf9\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.561051 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-credential-keys\") pod \"keystone-bootstrap-cwvf9\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.561100 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-combined-ca-bundle\") pod \"keystone-bootstrap-cwvf9\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.561153 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpnx6\" (UniqueName: \"kubernetes.io/projected/b8556431-841c-4372-8c93-fc3649d9b0f0-kube-api-access-wpnx6\") pod \"keystone-bootstrap-cwvf9\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.561200 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-scripts\") pod \"keystone-bootstrap-cwvf9\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.561222 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-config-data\") pod \"keystone-bootstrap-cwvf9\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.565988 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-fernet-keys\") pod \"keystone-bootstrap-cwvf9\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.566002 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-combined-ca-bundle\") pod \"keystone-bootstrap-cwvf9\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.566127 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-config-data\") pod \"keystone-bootstrap-cwvf9\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.566698 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-scripts\") pod \"keystone-bootstrap-cwvf9\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.576113 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-credential-keys\") pod \"keystone-bootstrap-cwvf9\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.576483 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpnx6\" (UniqueName: \"kubernetes.io/projected/b8556431-841c-4372-8c93-fc3649d9b0f0-kube-api-access-wpnx6\") pod \"keystone-bootstrap-cwvf9\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:18 crc kubenswrapper[4696]: I0321 08:47:18.707430 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:20 crc kubenswrapper[4696]: I0321 08:47:20.547410 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="782d4c3b-a1bf-4e02-ba5b-0274f885c26b" path="/var/lib/kubelet/pods/782d4c3b-a1bf-4e02-ba5b-0274f885c26b/volumes" Mar 21 08:47:22 crc kubenswrapper[4696]: I0321 08:47:22.180762 4696 scope.go:117] "RemoveContainer" containerID="7e7507b773e14ca3d089aa8075b93fec3216569eeeb079b66f4723631ab2cfc1" Mar 21 08:47:22 crc kubenswrapper[4696]: I0321 08:47:22.210233 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 08:47:24 crc kubenswrapper[4696]: I0321 08:47:24.561915 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"02fca1e0-3685-4d5b-86f9-bef0511876df","Type":"ContainerStarted","Data":"e660b83e06e17513aae66d0ffece19e7b8822ee6ee90e63e0664bbe24152e3b1"} Mar 21 08:47:26 crc kubenswrapper[4696]: I0321 08:47:26.580190 4696 generic.go:334] "Generic (PLEG): container finished" podID="c4ac0955-0d2a-4081-b82b-32dd288de552" containerID="f07398703df0547d265b0236bc994a5f3e2b860e2c22130759ce02432acf9cdf" exitCode=0 Mar 21 08:47:26 crc kubenswrapper[4696]: I0321 08:47:26.580534 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-m25wk" event={"ID":"c4ac0955-0d2a-4081-b82b-32dd288de552","Type":"ContainerDied","Data":"f07398703df0547d265b0236bc994a5f3e2b860e2c22130759ce02432acf9cdf"} Mar 21 08:47:30 crc kubenswrapper[4696]: I0321 08:47:30.340951 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:47:30 crc kubenswrapper[4696]: I0321 08:47:30.341514 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:47:35 crc kubenswrapper[4696]: I0321 08:47:35.119345 4696 scope.go:117] "RemoveContainer" containerID="97140fe2213ea9b227900bd9eb9d36aa99203d680f911ce09757b1a87aec226d" Mar 21 08:47:35 crc kubenswrapper[4696]: E0321 08:47:35.148283 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current" Mar 21 08:47:35 crc kubenswrapper[4696]: E0321 08:47:35.148333 4696 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current" Mar 21 08:47:35 crc kubenswrapper[4696]: E0321 08:47:35.148475 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cloudkitty-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CloudKittyPassword,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:CloudKittyPassword,Optional:nil,},},},EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:cloudkitty-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:certs,ReadOnly:true,MountPath:/var/lib/openstack/loki-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mv7dw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42406,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-db-sync-vgsz6_openstack(d3fc5362-53e5-4c35-be23-49dc3dd1fcba): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 08:47:35 crc kubenswrapper[4696]: E0321 08:47:35.149803 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cloudkitty-db-sync-vgsz6" podUID="d3fc5362-53e5-4c35-be23-49dc3dd1fcba" Mar 21 08:47:35 crc kubenswrapper[4696]: I0321 08:47:35.233871 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-m25wk" Mar 21 08:47:35 crc kubenswrapper[4696]: I0321 08:47:35.405437 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrw9c\" (UniqueName: \"kubernetes.io/projected/c4ac0955-0d2a-4081-b82b-32dd288de552-kube-api-access-xrw9c\") pod \"c4ac0955-0d2a-4081-b82b-32dd288de552\" (UID: \"c4ac0955-0d2a-4081-b82b-32dd288de552\") " Mar 21 08:47:35 crc kubenswrapper[4696]: I0321 08:47:35.405497 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c4ac0955-0d2a-4081-b82b-32dd288de552-config\") pod \"c4ac0955-0d2a-4081-b82b-32dd288de552\" (UID: \"c4ac0955-0d2a-4081-b82b-32dd288de552\") " Mar 21 08:47:35 crc kubenswrapper[4696]: I0321 08:47:35.405717 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4ac0955-0d2a-4081-b82b-32dd288de552-combined-ca-bundle\") pod \"c4ac0955-0d2a-4081-b82b-32dd288de552\" (UID: \"c4ac0955-0d2a-4081-b82b-32dd288de552\") " Mar 21 08:47:35 crc kubenswrapper[4696]: I0321 08:47:35.410347 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4ac0955-0d2a-4081-b82b-32dd288de552-kube-api-access-xrw9c" (OuterVolumeSpecName: "kube-api-access-xrw9c") pod "c4ac0955-0d2a-4081-b82b-32dd288de552" (UID: "c4ac0955-0d2a-4081-b82b-32dd288de552"). InnerVolumeSpecName "kube-api-access-xrw9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:47:35 crc kubenswrapper[4696]: I0321 08:47:35.430080 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4ac0955-0d2a-4081-b82b-32dd288de552-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c4ac0955-0d2a-4081-b82b-32dd288de552" (UID: "c4ac0955-0d2a-4081-b82b-32dd288de552"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:35 crc kubenswrapper[4696]: I0321 08:47:35.441157 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4ac0955-0d2a-4081-b82b-32dd288de552-config" (OuterVolumeSpecName: "config") pod "c4ac0955-0d2a-4081-b82b-32dd288de552" (UID: "c4ac0955-0d2a-4081-b82b-32dd288de552"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:35 crc kubenswrapper[4696]: I0321 08:47:35.507595 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4ac0955-0d2a-4081-b82b-32dd288de552-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:35 crc kubenswrapper[4696]: I0321 08:47:35.507635 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrw9c\" (UniqueName: \"kubernetes.io/projected/c4ac0955-0d2a-4081-b82b-32dd288de552-kube-api-access-xrw9c\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:35 crc kubenswrapper[4696]: I0321 08:47:35.507647 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c4ac0955-0d2a-4081-b82b-32dd288de552-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:35 crc kubenswrapper[4696]: I0321 08:47:35.686633 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-m25wk" event={"ID":"c4ac0955-0d2a-4081-b82b-32dd288de552","Type":"ContainerDied","Data":"f0d6a379964db7a6dd417f343874b928e2d85d3a88ed27102b05589c7808f438"} Mar 21 08:47:35 crc kubenswrapper[4696]: I0321 08:47:35.686680 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-m25wk" Mar 21 08:47:35 crc kubenswrapper[4696]: I0321 08:47:35.686692 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0d6a379964db7a6dd417f343874b928e2d85d3a88ed27102b05589c7808f438" Mar 21 08:47:35 crc kubenswrapper[4696]: E0321 08:47:35.688416 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current\\\"\"" pod="openstack/cloudkitty-db-sync-vgsz6" podUID="d3fc5362-53e5-4c35-be23-49dc3dd1fcba" Mar 21 08:47:36 crc kubenswrapper[4696]: E0321 08:47:36.322669 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Mar 21 08:47:36 crc kubenswrapper[4696]: E0321 08:47:36.323306 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dpdc5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-n5tdf_openstack(a70d1c8a-2db3-4ade-af61-fc8225262c18): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 08:47:36 crc kubenswrapper[4696]: E0321 08:47:36.325946 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-n5tdf" podUID="a70d1c8a-2db3-4ade-af61-fc8225262c18" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.688593 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-6mxl5"] Mar 21 08:47:36 crc kubenswrapper[4696]: E0321 08:47:36.689465 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ac0955-0d2a-4081-b82b-32dd288de552" containerName="neutron-db-sync" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.689540 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ac0955-0d2a-4081-b82b-32dd288de552" containerName="neutron-db-sync" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.689846 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ac0955-0d2a-4081-b82b-32dd288de552" containerName="neutron-db-sync" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.696349 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.709677 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-6mxl5"] Mar 21 08:47:36 crc kubenswrapper[4696]: E0321 08:47:36.750845 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-n5tdf" podUID="a70d1c8a-2db3-4ade-af61-fc8225262c18" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.849367 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-6mxl5\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.849416 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-config\") pod \"dnsmasq-dns-55f844cf75-6mxl5\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.849485 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-6mxl5\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.849504 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-6mxl5\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.849520 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-dns-svc\") pod \"dnsmasq-dns-55f844cf75-6mxl5\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.849540 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z98vc\" (UniqueName: \"kubernetes.io/projected/66c5da52-ed29-4222-8867-ac05aac6a382-kube-api-access-z98vc\") pod \"dnsmasq-dns-55f844cf75-6mxl5\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.901632 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5d77f68d84-8hrbc"] Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.903243 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d77f68d84-8hrbc" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.908042 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-9rtmn" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.908300 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.908349 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.908979 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.928994 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d77f68d84-8hrbc"] Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.951265 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-6mxl5\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.951309 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-6mxl5\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.951355 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-dns-svc\") pod \"dnsmasq-dns-55f844cf75-6mxl5\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.951377 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z98vc\" (UniqueName: \"kubernetes.io/projected/66c5da52-ed29-4222-8867-ac05aac6a382-kube-api-access-z98vc\") pod \"dnsmasq-dns-55f844cf75-6mxl5\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.953158 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-6mxl5\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.953195 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-config\") pod \"dnsmasq-dns-55f844cf75-6mxl5\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.954138 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-6mxl5\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.958515 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-dns-svc\") pod \"dnsmasq-dns-55f844cf75-6mxl5\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.958913 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-6mxl5\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.958940 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-config\") pod \"dnsmasq-dns-55f844cf75-6mxl5\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.966699 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-6mxl5\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:36 crc kubenswrapper[4696]: I0321 08:47:36.972946 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z98vc\" (UniqueName: \"kubernetes.io/projected/66c5da52-ed29-4222-8867-ac05aac6a382-kube-api-access-z98vc\") pod \"dnsmasq-dns-55f844cf75-6mxl5\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.057897 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwfxf\" (UniqueName: \"kubernetes.io/projected/9cba9c3e-f501-4351-b0af-26c07ff8c757-kube-api-access-vwfxf\") pod \"neutron-5d77f68d84-8hrbc\" (UID: \"9cba9c3e-f501-4351-b0af-26c07ff8c757\") " pod="openstack/neutron-5d77f68d84-8hrbc" Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.057979 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-httpd-config\") pod \"neutron-5d77f68d84-8hrbc\" (UID: \"9cba9c3e-f501-4351-b0af-26c07ff8c757\") " pod="openstack/neutron-5d77f68d84-8hrbc" Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.058050 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-ovndb-tls-certs\") pod \"neutron-5d77f68d84-8hrbc\" (UID: \"9cba9c3e-f501-4351-b0af-26c07ff8c757\") " pod="openstack/neutron-5d77f68d84-8hrbc" Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.058122 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-config\") pod \"neutron-5d77f68d84-8hrbc\" (UID: \"9cba9c3e-f501-4351-b0af-26c07ff8c757\") " pod="openstack/neutron-5d77f68d84-8hrbc" Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.058148 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-combined-ca-bundle\") pod \"neutron-5d77f68d84-8hrbc\" (UID: \"9cba9c3e-f501-4351-b0af-26c07ff8c757\") " pod="openstack/neutron-5d77f68d84-8hrbc" Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.094752 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-cwvf9"] Mar 21 08:47:37 crc kubenswrapper[4696]: W0321 08:47:37.107500 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8556431_841c_4372_8c93_fc3649d9b0f0.slice/crio-d5a79c621124d5b10c6d5d4fb06b3b837fb1ac3b3ab4da0e9f46f9fd9a0bb7d6 WatchSource:0}: Error finding container d5a79c621124d5b10c6d5d4fb06b3b837fb1ac3b3ab4da0e9f46f9fd9a0bb7d6: Status 404 returned error can't find the container with id d5a79c621124d5b10c6d5d4fb06b3b837fb1ac3b3ab4da0e9f46f9fd9a0bb7d6 Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.141126 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.159174 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-ovndb-tls-certs\") pod \"neutron-5d77f68d84-8hrbc\" (UID: \"9cba9c3e-f501-4351-b0af-26c07ff8c757\") " pod="openstack/neutron-5d77f68d84-8hrbc" Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.159232 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-config\") pod \"neutron-5d77f68d84-8hrbc\" (UID: \"9cba9c3e-f501-4351-b0af-26c07ff8c757\") " pod="openstack/neutron-5d77f68d84-8hrbc" Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.159255 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-combined-ca-bundle\") pod \"neutron-5d77f68d84-8hrbc\" (UID: \"9cba9c3e-f501-4351-b0af-26c07ff8c757\") " pod="openstack/neutron-5d77f68d84-8hrbc" Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.159305 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwfxf\" (UniqueName: \"kubernetes.io/projected/9cba9c3e-f501-4351-b0af-26c07ff8c757-kube-api-access-vwfxf\") pod \"neutron-5d77f68d84-8hrbc\" (UID: \"9cba9c3e-f501-4351-b0af-26c07ff8c757\") " pod="openstack/neutron-5d77f68d84-8hrbc" Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.159353 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-httpd-config\") pod \"neutron-5d77f68d84-8hrbc\" (UID: \"9cba9c3e-f501-4351-b0af-26c07ff8c757\") " pod="openstack/neutron-5d77f68d84-8hrbc" Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.163694 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-httpd-config\") pod \"neutron-5d77f68d84-8hrbc\" (UID: \"9cba9c3e-f501-4351-b0af-26c07ff8c757\") " pod="openstack/neutron-5d77f68d84-8hrbc" Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.170549 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-combined-ca-bundle\") pod \"neutron-5d77f68d84-8hrbc\" (UID: \"9cba9c3e-f501-4351-b0af-26c07ff8c757\") " pod="openstack/neutron-5d77f68d84-8hrbc" Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.172037 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-ovndb-tls-certs\") pod \"neutron-5d77f68d84-8hrbc\" (UID: \"9cba9c3e-f501-4351-b0af-26c07ff8c757\") " pod="openstack/neutron-5d77f68d84-8hrbc" Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.185201 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-config\") pod \"neutron-5d77f68d84-8hrbc\" (UID: \"9cba9c3e-f501-4351-b0af-26c07ff8c757\") " pod="openstack/neutron-5d77f68d84-8hrbc" Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.199622 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwfxf\" (UniqueName: \"kubernetes.io/projected/9cba9c3e-f501-4351-b0af-26c07ff8c757-kube-api-access-vwfxf\") pod \"neutron-5d77f68d84-8hrbc\" (UID: \"9cba9c3e-f501-4351-b0af-26c07ff8c757\") " pod="openstack/neutron-5d77f68d84-8hrbc" Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.238300 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d77f68d84-8hrbc" Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.762745 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbd625f3-5fcc-48f5-9022-df08fdc01887","Type":"ContainerStarted","Data":"7d3f92e8692b9fc0fa147cacc12efb0194a77da1acc10ad757b8fb7c002a9a62"} Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.773622 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nqvgp" event={"ID":"80364601-ba09-4192-a176-e04ad8d45506","Type":"ContainerStarted","Data":"47939aa8ea3147e4b4b4c629c0914bbb4bd10f5faf5fc4b3728caa3b9d4324cb"} Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.782608 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"296e9317-a507-4e57-8ee3-d42a091f866c","Type":"ContainerStarted","Data":"1c5e7e384295a1621158846edea9c769dbc91bd01ee2e2325ad8211be97cf9e6"} Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.818249 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-nqvgp" podStartSLOduration=17.460009278 podStartE2EDuration="35.818229314s" podCreationTimestamp="2026-03-21 08:47:02 +0000 UTC" firstStartedPulling="2026-03-21 08:47:03.798636019 +0000 UTC m=+1157.919516732" lastFinishedPulling="2026-03-21 08:47:22.156856055 +0000 UTC m=+1176.277736768" observedRunningTime="2026-03-21 08:47:37.799545292 +0000 UTC m=+1191.920426005" watchObservedRunningTime="2026-03-21 08:47:37.818229314 +0000 UTC m=+1191.939110027" Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.823114 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"02fca1e0-3685-4d5b-86f9-bef0511876df","Type":"ContainerStarted","Data":"d8217000d9f510c75a9c4ac25dd08ed01b612136d534442ece500e55d7f463dd"} Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.837660 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rzcz6" event={"ID":"4381bdfb-820a-42cc-9a0f-0ca3e6f34449","Type":"ContainerStarted","Data":"76e69cc985d933b3ea0095ce375e81e97f585398126f36881f496d0ff4681010"} Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.846066 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-cwvf9" event={"ID":"b8556431-841c-4372-8c93-fc3649d9b0f0","Type":"ContainerStarted","Data":"00da315452c455fa030a1e82cea1365333e9da4c1223f6f94970f47b1284776e"} Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.846112 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-cwvf9" event={"ID":"b8556431-841c-4372-8c93-fc3649d9b0f0","Type":"ContainerStarted","Data":"d5a79c621124d5b10c6d5d4fb06b3b837fb1ac3b3ab4da0e9f46f9fd9a0bb7d6"} Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.861590 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-rzcz6" podStartSLOduration=5.215816059 podStartE2EDuration="35.861569333s" podCreationTimestamp="2026-03-21 08:47:02 +0000 UTC" firstStartedPulling="2026-03-21 08:47:03.962780751 +0000 UTC m=+1158.083661464" lastFinishedPulling="2026-03-21 08:47:34.608534025 +0000 UTC m=+1188.729414738" observedRunningTime="2026-03-21 08:47:37.857319195 +0000 UTC m=+1191.978199908" watchObservedRunningTime="2026-03-21 08:47:37.861569333 +0000 UTC m=+1191.982450046" Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.893933 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-cwvf9" podStartSLOduration=19.893905516 podStartE2EDuration="19.893905516s" podCreationTimestamp="2026-03-21 08:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:47:37.8800905 +0000 UTC m=+1192.000971213" watchObservedRunningTime="2026-03-21 08:47:37.893905516 +0000 UTC m=+1192.014786239" Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.984963 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-6mxl5"] Mar 21 08:47:37 crc kubenswrapper[4696]: W0321 08:47:37.988799 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66c5da52_ed29_4222_8867_ac05aac6a382.slice/crio-c5a2b89c8b11dd92388e049e0363d4ef438dacf9e3bad408b15e6ae38124c69e WatchSource:0}: Error finding container c5a2b89c8b11dd92388e049e0363d4ef438dacf9e3bad408b15e6ae38124c69e: Status 404 returned error can't find the container with id c5a2b89c8b11dd92388e049e0363d4ef438dacf9e3bad408b15e6ae38124c69e Mar 21 08:47:37 crc kubenswrapper[4696]: I0321 08:47:37.996316 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d77f68d84-8hrbc"] Mar 21 08:47:38 crc kubenswrapper[4696]: I0321 08:47:38.859290 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d77f68d84-8hrbc" event={"ID":"9cba9c3e-f501-4351-b0af-26c07ff8c757","Type":"ContainerStarted","Data":"fe5af300da48a9658e1669e9924993a9cbf0038e20390138a155bf19acc5456a"} Mar 21 08:47:38 crc kubenswrapper[4696]: I0321 08:47:38.859931 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d77f68d84-8hrbc" event={"ID":"9cba9c3e-f501-4351-b0af-26c07ff8c757","Type":"ContainerStarted","Data":"540d30b2c60a2e402249c23a1cbbbcd811f4ad5d55f8b474e447c905ac1eb40f"} Mar 21 08:47:38 crc kubenswrapper[4696]: I0321 08:47:38.865470 4696 generic.go:334] "Generic (PLEG): container finished" podID="66c5da52-ed29-4222-8867-ac05aac6a382" containerID="cfffd805cedc9becdabd33fdd877d3f0ffbf1f5ac7d9f680703abc9d2b48c5a7" exitCode=0 Mar 21 08:47:38 crc kubenswrapper[4696]: I0321 08:47:38.865897 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" event={"ID":"66c5da52-ed29-4222-8867-ac05aac6a382","Type":"ContainerDied","Data":"cfffd805cedc9becdabd33fdd877d3f0ffbf1f5ac7d9f680703abc9d2b48c5a7"} Mar 21 08:47:38 crc kubenswrapper[4696]: I0321 08:47:38.865961 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" event={"ID":"66c5da52-ed29-4222-8867-ac05aac6a382","Type":"ContainerStarted","Data":"c5a2b89c8b11dd92388e049e0363d4ef438dacf9e3bad408b15e6ae38124c69e"} Mar 21 08:47:38 crc kubenswrapper[4696]: I0321 08:47:38.889070 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="296e9317-a507-4e57-8ee3-d42a091f866c" containerName="glance-log" containerID="cri-o://1c5e7e384295a1621158846edea9c769dbc91bd01ee2e2325ad8211be97cf9e6" gracePeriod=30 Mar 21 08:47:38 crc kubenswrapper[4696]: I0321 08:47:38.889254 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"296e9317-a507-4e57-8ee3-d42a091f866c","Type":"ContainerStarted","Data":"02f71b56389f4f65c6fb6ec54074e2adfea253d692cbac199e64991a0696ee68"} Mar 21 08:47:38 crc kubenswrapper[4696]: I0321 08:47:38.889369 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="296e9317-a507-4e57-8ee3-d42a091f866c" containerName="glance-httpd" containerID="cri-o://02f71b56389f4f65c6fb6ec54074e2adfea253d692cbac199e64991a0696ee68" gracePeriod=30 Mar 21 08:47:38 crc kubenswrapper[4696]: I0321 08:47:38.969342 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=29.969321606 podStartE2EDuration="29.969321606s" podCreationTimestamp="2026-03-21 08:47:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:47:38.935966145 +0000 UTC m=+1193.056846858" watchObservedRunningTime="2026-03-21 08:47:38.969321606 +0000 UTC m=+1193.090202319" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.446834 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-65c4b447c5-fstxp"] Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.449277 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.454557 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.454842 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.468488 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-65c4b447c5-fstxp"] Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.533447 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gj22\" (UniqueName: \"kubernetes.io/projected/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-kube-api-access-5gj22\") pod \"neutron-65c4b447c5-fstxp\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.533519 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-httpd-config\") pod \"neutron-65c4b447c5-fstxp\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.533565 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-public-tls-certs\") pod \"neutron-65c4b447c5-fstxp\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.533638 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-internal-tls-certs\") pod \"neutron-65c4b447c5-fstxp\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.533682 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-config\") pod \"neutron-65c4b447c5-fstxp\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.533714 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-ovndb-tls-certs\") pod \"neutron-65c4b447c5-fstxp\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.533758 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-combined-ca-bundle\") pod \"neutron-65c4b447c5-fstxp\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.635899 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-httpd-config\") pod \"neutron-65c4b447c5-fstxp\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.635960 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-public-tls-certs\") pod \"neutron-65c4b447c5-fstxp\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.636045 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-internal-tls-certs\") pod \"neutron-65c4b447c5-fstxp\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.636074 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-config\") pod \"neutron-65c4b447c5-fstxp\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.636118 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-ovndb-tls-certs\") pod \"neutron-65c4b447c5-fstxp\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.636152 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-combined-ca-bundle\") pod \"neutron-65c4b447c5-fstxp\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.636218 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gj22\" (UniqueName: \"kubernetes.io/projected/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-kube-api-access-5gj22\") pod \"neutron-65c4b447c5-fstxp\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.644136 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-config\") pod \"neutron-65c4b447c5-fstxp\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.644492 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-public-tls-certs\") pod \"neutron-65c4b447c5-fstxp\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.645868 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-combined-ca-bundle\") pod \"neutron-65c4b447c5-fstxp\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.647875 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-httpd-config\") pod \"neutron-65c4b447c5-fstxp\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.657802 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-ovndb-tls-certs\") pod \"neutron-65c4b447c5-fstxp\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.659066 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-internal-tls-certs\") pod \"neutron-65c4b447c5-fstxp\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.661528 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gj22\" (UniqueName: \"kubernetes.io/projected/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-kube-api-access-5gj22\") pod \"neutron-65c4b447c5-fstxp\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.796534 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.898684 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"02fca1e0-3685-4d5b-86f9-bef0511876df","Type":"ContainerStarted","Data":"da6a0ddb2bd8efcc6a441f66dfa4a515e7e3c637ea78a4a987348b985d86dd9d"} Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.901590 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d77f68d84-8hrbc" event={"ID":"9cba9c3e-f501-4351-b0af-26c07ff8c757","Type":"ContainerStarted","Data":"cb493038d7980c2cefcca5b04d8e6520ea16a047f9b766eef873b8ed1ece417d"} Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.902735 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5d77f68d84-8hrbc" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.905152 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" event={"ID":"66c5da52-ed29-4222-8867-ac05aac6a382","Type":"ContainerStarted","Data":"e26cba575b106923c7da050f84d362a3bb8475f3b104a4d2d5e7361c6fa84e77"} Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.905618 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.909365 4696 generic.go:334] "Generic (PLEG): container finished" podID="296e9317-a507-4e57-8ee3-d42a091f866c" containerID="02f71b56389f4f65c6fb6ec54074e2adfea253d692cbac199e64991a0696ee68" exitCode=0 Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.909398 4696 generic.go:334] "Generic (PLEG): container finished" podID="296e9317-a507-4e57-8ee3-d42a091f866c" containerID="1c5e7e384295a1621158846edea9c769dbc91bd01ee2e2325ad8211be97cf9e6" exitCode=143 Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.909399 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"296e9317-a507-4e57-8ee3-d42a091f866c","Type":"ContainerDied","Data":"02f71b56389f4f65c6fb6ec54074e2adfea253d692cbac199e64991a0696ee68"} Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.909444 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"296e9317-a507-4e57-8ee3-d42a091f866c","Type":"ContainerDied","Data":"1c5e7e384295a1621158846edea9c769dbc91bd01ee2e2325ad8211be97cf9e6"} Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.929587 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=27.929566962 podStartE2EDuration="27.929566962s" podCreationTimestamp="2026-03-21 08:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:47:39.92271022 +0000 UTC m=+1194.043590943" watchObservedRunningTime="2026-03-21 08:47:39.929566962 +0000 UTC m=+1194.050447675" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.929919 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.929959 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:39 crc kubenswrapper[4696]: I0321 08:47:39.965597 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5d77f68d84-8hrbc" podStartSLOduration=3.9655753259999997 podStartE2EDuration="3.965575326s" podCreationTimestamp="2026-03-21 08:47:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:47:39.953161739 +0000 UTC m=+1194.074042452" watchObservedRunningTime="2026-03-21 08:47:39.965575326 +0000 UTC m=+1194.086456039" Mar 21 08:47:40 crc kubenswrapper[4696]: I0321 08:47:40.001619 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" podStartSLOduration=4.001601962 podStartE2EDuration="4.001601962s" podCreationTimestamp="2026-03-21 08:47:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:47:39.999012879 +0000 UTC m=+1194.119893592" watchObservedRunningTime="2026-03-21 08:47:40.001601962 +0000 UTC m=+1194.122482675" Mar 21 08:47:40 crc kubenswrapper[4696]: I0321 08:47:40.803787 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-65c4b447c5-fstxp"] Mar 21 08:47:42 crc kubenswrapper[4696]: I0321 08:47:42.800867 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 21 08:47:42 crc kubenswrapper[4696]: I0321 08:47:42.802581 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 21 08:47:42 crc kubenswrapper[4696]: I0321 08:47:42.802656 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 21 08:47:42 crc kubenswrapper[4696]: I0321 08:47:42.802723 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 21 08:47:42 crc kubenswrapper[4696]: I0321 08:47:42.830215 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 21 08:47:42 crc kubenswrapper[4696]: I0321 08:47:42.838035 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 21 08:47:43 crc kubenswrapper[4696]: W0321 08:47:43.741172 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19b4b8b0_047b_48bf_9c9b_70fe162cf5bd.slice/crio-de19182eab2d7b11b3b6fc702051f49c65be082ba99d7df0443f6f1bf8947a4a WatchSource:0}: Error finding container de19182eab2d7b11b3b6fc702051f49c65be082ba99d7df0443f6f1bf8947a4a: Status 404 returned error can't find the container with id de19182eab2d7b11b3b6fc702051f49c65be082ba99d7df0443f6f1bf8947a4a Mar 21 08:47:43 crc kubenswrapper[4696]: I0321 08:47:43.974330 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65c4b447c5-fstxp" event={"ID":"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd","Type":"ContainerStarted","Data":"de19182eab2d7b11b3b6fc702051f49c65be082ba99d7df0443f6f1bf8947a4a"} Mar 21 08:47:43 crc kubenswrapper[4696]: I0321 08:47:43.976503 4696 generic.go:334] "Generic (PLEG): container finished" podID="80364601-ba09-4192-a176-e04ad8d45506" containerID="47939aa8ea3147e4b4b4c629c0914bbb4bd10f5faf5fc4b3728caa3b9d4324cb" exitCode=0 Mar 21 08:47:43 crc kubenswrapper[4696]: I0321 08:47:43.976558 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nqvgp" event={"ID":"80364601-ba09-4192-a176-e04ad8d45506","Type":"ContainerDied","Data":"47939aa8ea3147e4b4b4c629c0914bbb4bd10f5faf5fc4b3728caa3b9d4324cb"} Mar 21 08:47:43 crc kubenswrapper[4696]: I0321 08:47:43.980480 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"296e9317-a507-4e57-8ee3-d42a091f866c","Type":"ContainerDied","Data":"31dd2530f6846337a05986d4f3310cf13376e7e01a1160b15069c17b3fcb081a"} Mar 21 08:47:43 crc kubenswrapper[4696]: I0321 08:47:43.980521 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31dd2530f6846337a05986d4f3310cf13376e7e01a1160b15069c17b3fcb081a" Mar 21 08:47:43 crc kubenswrapper[4696]: I0321 08:47:43.988310 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.037309 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/296e9317-a507-4e57-8ee3-d42a091f866c-config-data\") pod \"296e9317-a507-4e57-8ee3-d42a091f866c\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.037381 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/296e9317-a507-4e57-8ee3-d42a091f866c-logs\") pod \"296e9317-a507-4e57-8ee3-d42a091f866c\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.037431 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzr66\" (UniqueName: \"kubernetes.io/projected/296e9317-a507-4e57-8ee3-d42a091f866c-kube-api-access-fzr66\") pod \"296e9317-a507-4e57-8ee3-d42a091f866c\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.037491 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/296e9317-a507-4e57-8ee3-d42a091f866c-httpd-run\") pod \"296e9317-a507-4e57-8ee3-d42a091f866c\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.037658 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") pod \"296e9317-a507-4e57-8ee3-d42a091f866c\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.037701 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/296e9317-a507-4e57-8ee3-d42a091f866c-combined-ca-bundle\") pod \"296e9317-a507-4e57-8ee3-d42a091f866c\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.037806 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/296e9317-a507-4e57-8ee3-d42a091f866c-scripts\") pod \"296e9317-a507-4e57-8ee3-d42a091f866c\" (UID: \"296e9317-a507-4e57-8ee3-d42a091f866c\") " Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.047173 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/296e9317-a507-4e57-8ee3-d42a091f866c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "296e9317-a507-4e57-8ee3-d42a091f866c" (UID: "296e9317-a507-4e57-8ee3-d42a091f866c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.047295 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/296e9317-a507-4e57-8ee3-d42a091f866c-kube-api-access-fzr66" (OuterVolumeSpecName: "kube-api-access-fzr66") pod "296e9317-a507-4e57-8ee3-d42a091f866c" (UID: "296e9317-a507-4e57-8ee3-d42a091f866c"). InnerVolumeSpecName "kube-api-access-fzr66". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.048710 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/296e9317-a507-4e57-8ee3-d42a091f866c-scripts" (OuterVolumeSpecName: "scripts") pod "296e9317-a507-4e57-8ee3-d42a091f866c" (UID: "296e9317-a507-4e57-8ee3-d42a091f866c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.049174 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/296e9317-a507-4e57-8ee3-d42a091f866c-logs" (OuterVolumeSpecName: "logs") pod "296e9317-a507-4e57-8ee3-d42a091f866c" (UID: "296e9317-a507-4e57-8ee3-d42a091f866c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.116538 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205" (OuterVolumeSpecName: "glance") pod "296e9317-a507-4e57-8ee3-d42a091f866c" (UID: "296e9317-a507-4e57-8ee3-d42a091f866c"). InnerVolumeSpecName "pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.124757 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/296e9317-a507-4e57-8ee3-d42a091f866c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "296e9317-a507-4e57-8ee3-d42a091f866c" (UID: "296e9317-a507-4e57-8ee3-d42a091f866c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.141048 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/296e9317-a507-4e57-8ee3-d42a091f866c-logs\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.141072 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzr66\" (UniqueName: \"kubernetes.io/projected/296e9317-a507-4e57-8ee3-d42a091f866c-kube-api-access-fzr66\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.141083 4696 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/296e9317-a507-4e57-8ee3-d42a091f866c-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.141112 4696 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") on node \"crc\" " Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.141123 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/296e9317-a507-4e57-8ee3-d42a091f866c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.141131 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/296e9317-a507-4e57-8ee3-d42a091f866c-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.171082 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/296e9317-a507-4e57-8ee3-d42a091f866c-config-data" (OuterVolumeSpecName: "config-data") pod "296e9317-a507-4e57-8ee3-d42a091f866c" (UID: "296e9317-a507-4e57-8ee3-d42a091f866c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.181369 4696 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.181515 4696 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205") on node "crc" Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.243446 4696 reconciler_common.go:293] "Volume detached for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.243479 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/296e9317-a507-4e57-8ee3-d42a091f866c-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:44 crc kubenswrapper[4696]: E0321 08:47:44.264306 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod80364601_ba09_4192_a176_e04ad8d45506.slice/crio-conmon-47939aa8ea3147e4b4b4c629c0914bbb4bd10f5faf5fc4b3728caa3b9d4324cb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod80364601_ba09_4192_a176_e04ad8d45506.slice/crio-47939aa8ea3147e4b4b4c629c0914bbb4bd10f5faf5fc4b3728caa3b9d4324cb.scope\": RecentStats: unable to find data in memory cache]" Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.990400 4696 generic.go:334] "Generic (PLEG): container finished" podID="4381bdfb-820a-42cc-9a0f-0ca3e6f34449" containerID="76e69cc985d933b3ea0095ce375e81e97f585398126f36881f496d0ff4681010" exitCode=0 Mar 21 08:47:44 crc kubenswrapper[4696]: I0321 08:47:44.990664 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rzcz6" event={"ID":"4381bdfb-820a-42cc-9a0f-0ca3e6f34449","Type":"ContainerDied","Data":"76e69cc985d933b3ea0095ce375e81e97f585398126f36881f496d0ff4681010"} Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.002989 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65c4b447c5-fstxp" event={"ID":"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd","Type":"ContainerStarted","Data":"5f83ec04b39227800d19bc750797e0e8b8ffaaee0a89e3f3298d96b0a49de970"} Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.003029 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65c4b447c5-fstxp" event={"ID":"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd","Type":"ContainerStarted","Data":"34604a8d20dd45769c938fc4e612566bd5eb98466fd47a1560d6276fec96a001"} Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.003139 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.006474 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbd625f3-5fcc-48f5-9022-df08fdc01887","Type":"ContainerStarted","Data":"0c5965d091d4ecbb20a816b1558e00f85fbc6379914b0c2e493d1c70c053bb4d"} Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.014957 4696 generic.go:334] "Generic (PLEG): container finished" podID="b8556431-841c-4372-8c93-fc3649d9b0f0" containerID="00da315452c455fa030a1e82cea1365333e9da4c1223f6f94970f47b1284776e" exitCode=0 Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.015099 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-cwvf9" event={"ID":"b8556431-841c-4372-8c93-fc3649d9b0f0","Type":"ContainerDied","Data":"00da315452c455fa030a1e82cea1365333e9da4c1223f6f94970f47b1284776e"} Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.015221 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.062659 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-65c4b447c5-fstxp" podStartSLOduration=6.062638352 podStartE2EDuration="6.062638352s" podCreationTimestamp="2026-03-21 08:47:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:47:45.050903644 +0000 UTC m=+1199.171784367" watchObservedRunningTime="2026-03-21 08:47:45.062638352 +0000 UTC m=+1199.183519065" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.100861 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.126805 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.155880 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.167233 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 08:47:45 crc kubenswrapper[4696]: E0321 08:47:45.167744 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="296e9317-a507-4e57-8ee3-d42a091f866c" containerName="glance-log" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.167759 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="296e9317-a507-4e57-8ee3-d42a091f866c" containerName="glance-log" Mar 21 08:47:45 crc kubenswrapper[4696]: E0321 08:47:45.167802 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="296e9317-a507-4e57-8ee3-d42a091f866c" containerName="glance-httpd" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.167809 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="296e9317-a507-4e57-8ee3-d42a091f866c" containerName="glance-httpd" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.168040 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="296e9317-a507-4e57-8ee3-d42a091f866c" containerName="glance-log" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.168060 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="296e9317-a507-4e57-8ee3-d42a091f866c" containerName="glance-httpd" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.169314 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.174004 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.174187 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.178301 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.296076 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.296456 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.296600 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.296744 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6n56\" (UniqueName: \"kubernetes.io/projected/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-kube-api-access-v6n56\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.296852 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.296982 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-logs\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.297094 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.297186 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.400509 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.400671 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.400766 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.400917 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6n56\" (UniqueName: \"kubernetes.io/projected/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-kube-api-access-v6n56\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.401054 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.401150 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-logs\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.401230 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.401284 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.409638 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-logs\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.410203 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.417274 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.418567 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.424549 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.425828 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.437041 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6n56\" (UniqueName: \"kubernetes.io/projected/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-kube-api-access-v6n56\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.438272 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.438320 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/36feb21b6af5aad5e90205d24e3f1d239a0e9aaaadf2ca8cf7f1809b63388e1a/globalmount\"" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.524772 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") pod \"glance-default-internal-api-0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.590778 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nqvgp" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.709844 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trrnn\" (UniqueName: \"kubernetes.io/projected/80364601-ba09-4192-a176-e04ad8d45506-kube-api-access-trrnn\") pod \"80364601-ba09-4192-a176-e04ad8d45506\" (UID: \"80364601-ba09-4192-a176-e04ad8d45506\") " Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.709928 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80364601-ba09-4192-a176-e04ad8d45506-scripts\") pod \"80364601-ba09-4192-a176-e04ad8d45506\" (UID: \"80364601-ba09-4192-a176-e04ad8d45506\") " Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.709955 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80364601-ba09-4192-a176-e04ad8d45506-combined-ca-bundle\") pod \"80364601-ba09-4192-a176-e04ad8d45506\" (UID: \"80364601-ba09-4192-a176-e04ad8d45506\") " Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.710030 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80364601-ba09-4192-a176-e04ad8d45506-config-data\") pod \"80364601-ba09-4192-a176-e04ad8d45506\" (UID: \"80364601-ba09-4192-a176-e04ad8d45506\") " Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.710057 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80364601-ba09-4192-a176-e04ad8d45506-logs\") pod \"80364601-ba09-4192-a176-e04ad8d45506\" (UID: \"80364601-ba09-4192-a176-e04ad8d45506\") " Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.712478 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80364601-ba09-4192-a176-e04ad8d45506-logs" (OuterVolumeSpecName: "logs") pod "80364601-ba09-4192-a176-e04ad8d45506" (UID: "80364601-ba09-4192-a176-e04ad8d45506"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.734787 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80364601-ba09-4192-a176-e04ad8d45506-kube-api-access-trrnn" (OuterVolumeSpecName: "kube-api-access-trrnn") pod "80364601-ba09-4192-a176-e04ad8d45506" (UID: "80364601-ba09-4192-a176-e04ad8d45506"). InnerVolumeSpecName "kube-api-access-trrnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.734907 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80364601-ba09-4192-a176-e04ad8d45506-scripts" (OuterVolumeSpecName: "scripts") pod "80364601-ba09-4192-a176-e04ad8d45506" (UID: "80364601-ba09-4192-a176-e04ad8d45506"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.782377 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80364601-ba09-4192-a176-e04ad8d45506-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "80364601-ba09-4192-a176-e04ad8d45506" (UID: "80364601-ba09-4192-a176-e04ad8d45506"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.785595 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80364601-ba09-4192-a176-e04ad8d45506-config-data" (OuterVolumeSpecName: "config-data") pod "80364601-ba09-4192-a176-e04ad8d45506" (UID: "80364601-ba09-4192-a176-e04ad8d45506"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.792137 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.828019 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trrnn\" (UniqueName: \"kubernetes.io/projected/80364601-ba09-4192-a176-e04ad8d45506-kube-api-access-trrnn\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.828050 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80364601-ba09-4192-a176-e04ad8d45506-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.828060 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80364601-ba09-4192-a176-e04ad8d45506-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.828069 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80364601-ba09-4192-a176-e04ad8d45506-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:45 crc kubenswrapper[4696]: I0321 08:47:45.828077 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80364601-ba09-4192-a176-e04ad8d45506-logs\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.052674 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nqvgp" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.054017 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nqvgp" event={"ID":"80364601-ba09-4192-a176-e04ad8d45506","Type":"ContainerDied","Data":"34f92239a190f120cdae216575bf329d8b3b5c80eb8a4139c0367ba3daa6b441"} Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.054176 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34f92239a190f120cdae216575bf329d8b3b5c80eb8a4139c0367ba3daa6b441" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.222425 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-565c47bc8-bmdxv"] Mar 21 08:47:46 crc kubenswrapper[4696]: E0321 08:47:46.223155 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80364601-ba09-4192-a176-e04ad8d45506" containerName="placement-db-sync" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.223171 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="80364601-ba09-4192-a176-e04ad8d45506" containerName="placement-db-sync" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.223401 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="80364601-ba09-4192-a176-e04ad8d45506" containerName="placement-db-sync" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.229987 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.235516 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-565c47bc8-bmdxv"] Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.238943 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.239103 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-f2wb5" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.239244 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.239341 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.239488 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.340798 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-scripts\") pod \"placement-565c47bc8-bmdxv\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.340857 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk4pp\" (UniqueName: \"kubernetes.io/projected/a32ac807-7955-425a-a971-e07ddde36bb0-kube-api-access-xk4pp\") pod \"placement-565c47bc8-bmdxv\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.340913 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-public-tls-certs\") pod \"placement-565c47bc8-bmdxv\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.341004 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-config-data\") pod \"placement-565c47bc8-bmdxv\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.341035 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-combined-ca-bundle\") pod \"placement-565c47bc8-bmdxv\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.341052 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a32ac807-7955-425a-a971-e07ddde36bb0-logs\") pod \"placement-565c47bc8-bmdxv\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.341098 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-internal-tls-certs\") pod \"placement-565c47bc8-bmdxv\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.442406 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-internal-tls-certs\") pod \"placement-565c47bc8-bmdxv\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.442466 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-scripts\") pod \"placement-565c47bc8-bmdxv\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.442493 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk4pp\" (UniqueName: \"kubernetes.io/projected/a32ac807-7955-425a-a971-e07ddde36bb0-kube-api-access-xk4pp\") pod \"placement-565c47bc8-bmdxv\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.442522 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-public-tls-certs\") pod \"placement-565c47bc8-bmdxv\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.442606 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-config-data\") pod \"placement-565c47bc8-bmdxv\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.442636 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-combined-ca-bundle\") pod \"placement-565c47bc8-bmdxv\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.442649 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a32ac807-7955-425a-a971-e07ddde36bb0-logs\") pod \"placement-565c47bc8-bmdxv\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.443051 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a32ac807-7955-425a-a971-e07ddde36bb0-logs\") pod \"placement-565c47bc8-bmdxv\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.451033 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-internal-tls-certs\") pod \"placement-565c47bc8-bmdxv\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.451264 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-scripts\") pod \"placement-565c47bc8-bmdxv\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.456006 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-config-data\") pod \"placement-565c47bc8-bmdxv\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.457466 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-combined-ca-bundle\") pod \"placement-565c47bc8-bmdxv\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.466765 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-public-tls-certs\") pod \"placement-565c47bc8-bmdxv\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.473633 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk4pp\" (UniqueName: \"kubernetes.io/projected/a32ac807-7955-425a-a971-e07ddde36bb0-kube-api-access-xk4pp\") pod \"placement-565c47bc8-bmdxv\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.575330 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-f2wb5" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.584776 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.589508 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="296e9317-a507-4e57-8ee3-d42a091f866c" path="/var/lib/kubelet/pods/296e9317-a507-4e57-8ee3-d42a091f866c/volumes" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.783148 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.962093 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-credential-keys\") pod \"b8556431-841c-4372-8c93-fc3649d9b0f0\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.962405 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-scripts\") pod \"b8556431-841c-4372-8c93-fc3649d9b0f0\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.962453 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-fernet-keys\") pod \"b8556431-841c-4372-8c93-fc3649d9b0f0\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.962479 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-combined-ca-bundle\") pod \"b8556431-841c-4372-8c93-fc3649d9b0f0\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.962511 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-config-data\") pod \"b8556431-841c-4372-8c93-fc3649d9b0f0\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.962621 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpnx6\" (UniqueName: \"kubernetes.io/projected/b8556431-841c-4372-8c93-fc3649d9b0f0-kube-api-access-wpnx6\") pod \"b8556431-841c-4372-8c93-fc3649d9b0f0\" (UID: \"b8556431-841c-4372-8c93-fc3649d9b0f0\") " Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.965989 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.977627 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8556431-841c-4372-8c93-fc3649d9b0f0-kube-api-access-wpnx6" (OuterVolumeSpecName: "kube-api-access-wpnx6") pod "b8556431-841c-4372-8c93-fc3649d9b0f0" (UID: "b8556431-841c-4372-8c93-fc3649d9b0f0"). InnerVolumeSpecName "kube-api-access-wpnx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.979457 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-scripts" (OuterVolumeSpecName: "scripts") pod "b8556431-841c-4372-8c93-fc3649d9b0f0" (UID: "b8556431-841c-4372-8c93-fc3649d9b0f0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:46 crc kubenswrapper[4696]: I0321 08:47:46.981923 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b8556431-841c-4372-8c93-fc3649d9b0f0" (UID: "b8556431-841c-4372-8c93-fc3649d9b0f0"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:46.995805 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b8556431-841c-4372-8c93-fc3649d9b0f0" (UID: "b8556431-841c-4372-8c93-fc3649d9b0f0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.054469 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b8556431-841c-4372-8c93-fc3649d9b0f0" (UID: "b8556431-841c-4372-8c93-fc3649d9b0f0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.065353 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpnx6\" (UniqueName: \"kubernetes.io/projected/b8556431-841c-4372-8c93-fc3649d9b0f0-kube-api-access-wpnx6\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.065388 4696 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.065400 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.065409 4696 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.065419 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.069481 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1af36c17-5501-4d4e-87a9-e9b43cb9eef0","Type":"ContainerStarted","Data":"3a9bab0c350025ca189e82a6631725e712e1e13767d1bddd512764b878f9a05e"} Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.076546 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rzcz6" event={"ID":"4381bdfb-820a-42cc-9a0f-0ca3e6f34449","Type":"ContainerDied","Data":"38e1b22cb3a6ff8a04a90b85b6e070a3e96f95b86b0eeb9dbe2d46b4b0773b3d"} Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.076589 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38e1b22cb3a6ff8a04a90b85b6e070a3e96f95b86b0eeb9dbe2d46b4b0773b3d" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.097458 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-cwvf9" event={"ID":"b8556431-841c-4372-8c93-fc3649d9b0f0","Type":"ContainerDied","Data":"d5a79c621124d5b10c6d5d4fb06b3b837fb1ac3b3ab4da0e9f46f9fd9a0bb7d6"} Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.097512 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5a79c621124d5b10c6d5d4fb06b3b837fb1ac3b3ab4da0e9f46f9fd9a0bb7d6" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.097568 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-cwvf9" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.113180 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-config-data" (OuterVolumeSpecName: "config-data") pod "b8556431-841c-4372-8c93-fc3649d9b0f0" (UID: "b8556431-841c-4372-8c93-fc3649d9b0f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.144339 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.170359 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8556431-841c-4372-8c93-fc3649d9b0f0-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.254651 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.273909 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rzcz6" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.326928 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-696f547b5d-r9gcd"] Mar 21 08:47:47 crc kubenswrapper[4696]: E0321 08:47:47.327428 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8556431-841c-4372-8c93-fc3649d9b0f0" containerName="keystone-bootstrap" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.327446 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8556431-841c-4372-8c93-fc3649d9b0f0" containerName="keystone-bootstrap" Mar 21 08:47:47 crc kubenswrapper[4696]: E0321 08:47:47.327486 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4381bdfb-820a-42cc-9a0f-0ca3e6f34449" containerName="barbican-db-sync" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.327494 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4381bdfb-820a-42cc-9a0f-0ca3e6f34449" containerName="barbican-db-sync" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.327742 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="4381bdfb-820a-42cc-9a0f-0ca3e6f34449" containerName="barbican-db-sync" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.327756 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8556431-841c-4372-8c93-fc3649d9b0f0" containerName="keystone-bootstrap" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.328560 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.332903 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.332994 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.379494 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4381bdfb-820a-42cc-9a0f-0ca3e6f34449-combined-ca-bundle\") pod \"4381bdfb-820a-42cc-9a0f-0ca3e6f34449\" (UID: \"4381bdfb-820a-42cc-9a0f-0ca3e6f34449\") " Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.379604 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4381bdfb-820a-42cc-9a0f-0ca3e6f34449-db-sync-config-data\") pod \"4381bdfb-820a-42cc-9a0f-0ca3e6f34449\" (UID: \"4381bdfb-820a-42cc-9a0f-0ca3e6f34449\") " Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.379897 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stvs7\" (UniqueName: \"kubernetes.io/projected/4381bdfb-820a-42cc-9a0f-0ca3e6f34449-kube-api-access-stvs7\") pod \"4381bdfb-820a-42cc-9a0f-0ca3e6f34449\" (UID: \"4381bdfb-820a-42cc-9a0f-0ca3e6f34449\") " Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.395937 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4381bdfb-820a-42cc-9a0f-0ca3e6f34449-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4381bdfb-820a-42cc-9a0f-0ca3e6f34449" (UID: "4381bdfb-820a-42cc-9a0f-0ca3e6f34449"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.399119 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4381bdfb-820a-42cc-9a0f-0ca3e6f34449-kube-api-access-stvs7" (OuterVolumeSpecName: "kube-api-access-stvs7") pod "4381bdfb-820a-42cc-9a0f-0ca3e6f34449" (UID: "4381bdfb-820a-42cc-9a0f-0ca3e6f34449"). InnerVolumeSpecName "kube-api-access-stvs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.411748 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4381bdfb-820a-42cc-9a0f-0ca3e6f34449-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4381bdfb-820a-42cc-9a0f-0ca3e6f34449" (UID: "4381bdfb-820a-42cc-9a0f-0ca3e6f34449"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.412252 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-mddgz"] Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.413621 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" podUID="64b14a78-860a-48c3-ae69-16cc36c77be8" containerName="dnsmasq-dns" containerID="cri-o://17745124f29da5ac840d31851871596d27883037e3f61a184b8f34cb87fd4a9b" gracePeriod=10 Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.431205 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-696f547b5d-r9gcd"] Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.479108 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-565c47bc8-bmdxv"] Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.481496 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31091b53-3bf3-4704-96bc-c5c33eb55218-combined-ca-bundle\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.481558 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31091b53-3bf3-4704-96bc-c5c33eb55218-scripts\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.481628 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31091b53-3bf3-4704-96bc-c5c33eb55218-internal-tls-certs\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.481688 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/31091b53-3bf3-4704-96bc-c5c33eb55218-fernet-keys\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.481704 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31091b53-3bf3-4704-96bc-c5c33eb55218-public-tls-certs\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.481720 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/31091b53-3bf3-4704-96bc-c5c33eb55218-credential-keys\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.481745 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31091b53-3bf3-4704-96bc-c5c33eb55218-config-data\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.481773 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f27w\" (UniqueName: \"kubernetes.io/projected/31091b53-3bf3-4704-96bc-c5c33eb55218-kube-api-access-9f27w\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.481854 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stvs7\" (UniqueName: \"kubernetes.io/projected/4381bdfb-820a-42cc-9a0f-0ca3e6f34449-kube-api-access-stvs7\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.481865 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4381bdfb-820a-42cc-9a0f-0ca3e6f34449-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.481906 4696 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4381bdfb-820a-42cc-9a0f-0ca3e6f34449-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.583063 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31091b53-3bf3-4704-96bc-c5c33eb55218-internal-tls-certs\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.583130 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31091b53-3bf3-4704-96bc-c5c33eb55218-public-tls-certs\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.583148 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/31091b53-3bf3-4704-96bc-c5c33eb55218-fernet-keys\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.583163 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/31091b53-3bf3-4704-96bc-c5c33eb55218-credential-keys\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.583185 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31091b53-3bf3-4704-96bc-c5c33eb55218-config-data\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.583213 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f27w\" (UniqueName: \"kubernetes.io/projected/31091b53-3bf3-4704-96bc-c5c33eb55218-kube-api-access-9f27w\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.583257 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31091b53-3bf3-4704-96bc-c5c33eb55218-combined-ca-bundle\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.583865 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31091b53-3bf3-4704-96bc-c5c33eb55218-scripts\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.595336 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31091b53-3bf3-4704-96bc-c5c33eb55218-public-tls-certs\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.598836 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31091b53-3bf3-4704-96bc-c5c33eb55218-config-data\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.603447 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31091b53-3bf3-4704-96bc-c5c33eb55218-combined-ca-bundle\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.603504 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f27w\" (UniqueName: \"kubernetes.io/projected/31091b53-3bf3-4704-96bc-c5c33eb55218-kube-api-access-9f27w\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.605001 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/31091b53-3bf3-4704-96bc-c5c33eb55218-fernet-keys\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.605092 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31091b53-3bf3-4704-96bc-c5c33eb55218-scripts\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.605440 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/31091b53-3bf3-4704-96bc-c5c33eb55218-credential-keys\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.605668 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31091b53-3bf3-4704-96bc-c5c33eb55218-internal-tls-certs\") pod \"keystone-696f547b5d-r9gcd\" (UID: \"31091b53-3bf3-4704-96bc-c5c33eb55218\") " pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:47 crc kubenswrapper[4696]: I0321 08:47:47.832561 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.136144 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-vgsz6" event={"ID":"d3fc5362-53e5-4c35-be23-49dc3dd1fcba","Type":"ContainerStarted","Data":"83e39ef69714631ade43ac7b3e551d20a5d282e9d803ae8961e6ae213c1e452e"} Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.150504 4696 generic.go:334] "Generic (PLEG): container finished" podID="64b14a78-860a-48c3-ae69-16cc36c77be8" containerID="17745124f29da5ac840d31851871596d27883037e3f61a184b8f34cb87fd4a9b" exitCode=0 Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.150567 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" event={"ID":"64b14a78-860a-48c3-ae69-16cc36c77be8","Type":"ContainerDied","Data":"17745124f29da5ac840d31851871596d27883037e3f61a184b8f34cb87fd4a9b"} Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.150604 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" event={"ID":"64b14a78-860a-48c3-ae69-16cc36c77be8","Type":"ContainerDied","Data":"037087306e9eaf6637464e8aaa5a52f1a30f4bd4e5326e7b9bb91a6a781071e8"} Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.150614 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="037087306e9eaf6637464e8aaa5a52f1a30f4bd4e5326e7b9bb91a6a781071e8" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.163654 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rzcz6" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.164103 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-565c47bc8-bmdxv" event={"ID":"a32ac807-7955-425a-a971-e07ddde36bb0","Type":"ContainerStarted","Data":"55862ad01a19e70bcd5fe3ea21ef73df613a89f8d76d7c0d42d3c99ae1f5d4db"} Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.164134 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-565c47bc8-bmdxv" event={"ID":"a32ac807-7955-425a-a971-e07ddde36bb0","Type":"ContainerStarted","Data":"5c2afb4660e62cf6acd556874a3e10004cdcf9eb8e4a65657b808244452e0d04"} Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.181287 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-db-sync-vgsz6" podStartSLOduration=3.313572114 podStartE2EDuration="46.181263749s" podCreationTimestamp="2026-03-21 08:47:02 +0000 UTC" firstStartedPulling="2026-03-21 08:47:04.456016514 +0000 UTC m=+1158.576897227" lastFinishedPulling="2026-03-21 08:47:47.323708149 +0000 UTC m=+1201.444588862" observedRunningTime="2026-03-21 08:47:48.153159025 +0000 UTC m=+1202.274039738" watchObservedRunningTime="2026-03-21 08:47:48.181263749 +0000 UTC m=+1202.302144462" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.188878 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.302018 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-ovsdbserver-nb\") pod \"64b14a78-860a-48c3-ae69-16cc36c77be8\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.302495 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frckn\" (UniqueName: \"kubernetes.io/projected/64b14a78-860a-48c3-ae69-16cc36c77be8-kube-api-access-frckn\") pod \"64b14a78-860a-48c3-ae69-16cc36c77be8\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.302544 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-config\") pod \"64b14a78-860a-48c3-ae69-16cc36c77be8\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.302598 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-ovsdbserver-sb\") pod \"64b14a78-860a-48c3-ae69-16cc36c77be8\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.302617 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-dns-svc\") pod \"64b14a78-860a-48c3-ae69-16cc36c77be8\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.302660 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-dns-swift-storage-0\") pod \"64b14a78-860a-48c3-ae69-16cc36c77be8\" (UID: \"64b14a78-860a-48c3-ae69-16cc36c77be8\") " Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.308200 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64b14a78-860a-48c3-ae69-16cc36c77be8-kube-api-access-frckn" (OuterVolumeSpecName: "kube-api-access-frckn") pod "64b14a78-860a-48c3-ae69-16cc36c77be8" (UID: "64b14a78-860a-48c3-ae69-16cc36c77be8"). InnerVolumeSpecName "kube-api-access-frckn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.408286 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frckn\" (UniqueName: \"kubernetes.io/projected/64b14a78-860a-48c3-ae69-16cc36c77be8-kube-api-access-frckn\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.565163 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-756466fd6c-6vdmf"] Mar 21 08:47:48 crc kubenswrapper[4696]: E0321 08:47:48.565717 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64b14a78-860a-48c3-ae69-16cc36c77be8" containerName="init" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.565737 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="64b14a78-860a-48c3-ae69-16cc36c77be8" containerName="init" Mar 21 08:47:48 crc kubenswrapper[4696]: E0321 08:47:48.565765 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64b14a78-860a-48c3-ae69-16cc36c77be8" containerName="dnsmasq-dns" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.565774 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="64b14a78-860a-48c3-ae69-16cc36c77be8" containerName="dnsmasq-dns" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.566027 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="64b14a78-860a-48c3-ae69-16cc36c77be8" containerName="dnsmasq-dns" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.567987 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-756466fd6c-6vdmf" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.579245 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-6hxd9" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.585052 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-756466fd6c-6vdmf"] Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.592910 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.592992 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.648906 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7c894bbd86-9phxx"] Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.651042 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7c894bbd86-9phxx" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.662335 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.716351 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "64b14a78-860a-48c3-ae69-16cc36c77be8" (UID: "64b14a78-860a-48c3-ae69-16cc36c77be8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.721869 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-696f547b5d-r9gcd"] Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.723024 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/639973de-e4aa-46d4-817a-f9c410238b45-config-data\") pod \"barbican-worker-756466fd6c-6vdmf\" (UID: \"639973de-e4aa-46d4-817a-f9c410238b45\") " pod="openstack/barbican-worker-756466fd6c-6vdmf" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.723088 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/639973de-e4aa-46d4-817a-f9c410238b45-logs\") pod \"barbican-worker-756466fd6c-6vdmf\" (UID: \"639973de-e4aa-46d4-817a-f9c410238b45\") " pod="openstack/barbican-worker-756466fd6c-6vdmf" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.723131 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/639973de-e4aa-46d4-817a-f9c410238b45-config-data-custom\") pod \"barbican-worker-756466fd6c-6vdmf\" (UID: \"639973de-e4aa-46d4-817a-f9c410238b45\") " pod="openstack/barbican-worker-756466fd6c-6vdmf" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.723283 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/639973de-e4aa-46d4-817a-f9c410238b45-combined-ca-bundle\") pod \"barbican-worker-756466fd6c-6vdmf\" (UID: \"639973de-e4aa-46d4-817a-f9c410238b45\") " pod="openstack/barbican-worker-756466fd6c-6vdmf" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.723357 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgjbc\" (UniqueName: \"kubernetes.io/projected/639973de-e4aa-46d4-817a-f9c410238b45-kube-api-access-zgjbc\") pod \"barbican-worker-756466fd6c-6vdmf\" (UID: \"639973de-e4aa-46d4-817a-f9c410238b45\") " pod="openstack/barbican-worker-756466fd6c-6vdmf" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.723602 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.749304 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7c894bbd86-9phxx"] Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.779092 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-6jw7g"] Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.785359 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.838791 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adeff71a-2b97-46e9-a945-d69cf06e34d9-config-data\") pod \"barbican-keystone-listener-7c894bbd86-9phxx\" (UID: \"adeff71a-2b97-46e9-a945-d69cf06e34d9\") " pod="openstack/barbican-keystone-listener-7c894bbd86-9phxx" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.868791 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjjrc\" (UniqueName: \"kubernetes.io/projected/adeff71a-2b97-46e9-a945-d69cf06e34d9-kube-api-access-zjjrc\") pod \"barbican-keystone-listener-7c894bbd86-9phxx\" (UID: \"adeff71a-2b97-46e9-a945-d69cf06e34d9\") " pod="openstack/barbican-keystone-listener-7c894bbd86-9phxx" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.868918 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/639973de-e4aa-46d4-817a-f9c410238b45-combined-ca-bundle\") pod \"barbican-worker-756466fd6c-6vdmf\" (UID: \"639973de-e4aa-46d4-817a-f9c410238b45\") " pod="openstack/barbican-worker-756466fd6c-6vdmf" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.868986 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgjbc\" (UniqueName: \"kubernetes.io/projected/639973de-e4aa-46d4-817a-f9c410238b45-kube-api-access-zgjbc\") pod \"barbican-worker-756466fd6c-6vdmf\" (UID: \"639973de-e4aa-46d4-817a-f9c410238b45\") " pod="openstack/barbican-worker-756466fd6c-6vdmf" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.869055 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adeff71a-2b97-46e9-a945-d69cf06e34d9-combined-ca-bundle\") pod \"barbican-keystone-listener-7c894bbd86-9phxx\" (UID: \"adeff71a-2b97-46e9-a945-d69cf06e34d9\") " pod="openstack/barbican-keystone-listener-7c894bbd86-9phxx" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.869207 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/639973de-e4aa-46d4-817a-f9c410238b45-config-data\") pod \"barbican-worker-756466fd6c-6vdmf\" (UID: \"639973de-e4aa-46d4-817a-f9c410238b45\") " pod="openstack/barbican-worker-756466fd6c-6vdmf" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.869253 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/639973de-e4aa-46d4-817a-f9c410238b45-logs\") pod \"barbican-worker-756466fd6c-6vdmf\" (UID: \"639973de-e4aa-46d4-817a-f9c410238b45\") " pod="openstack/barbican-worker-756466fd6c-6vdmf" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.869304 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/639973de-e4aa-46d4-817a-f9c410238b45-config-data-custom\") pod \"barbican-worker-756466fd6c-6vdmf\" (UID: \"639973de-e4aa-46d4-817a-f9c410238b45\") " pod="openstack/barbican-worker-756466fd6c-6vdmf" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.869345 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adeff71a-2b97-46e9-a945-d69cf06e34d9-logs\") pod \"barbican-keystone-listener-7c894bbd86-9phxx\" (UID: \"adeff71a-2b97-46e9-a945-d69cf06e34d9\") " pod="openstack/barbican-keystone-listener-7c894bbd86-9phxx" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.869392 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/adeff71a-2b97-46e9-a945-d69cf06e34d9-config-data-custom\") pod \"barbican-keystone-listener-7c894bbd86-9phxx\" (UID: \"adeff71a-2b97-46e9-a945-d69cf06e34d9\") " pod="openstack/barbican-keystone-listener-7c894bbd86-9phxx" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.880760 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/639973de-e4aa-46d4-817a-f9c410238b45-logs\") pod \"barbican-worker-756466fd6c-6vdmf\" (UID: \"639973de-e4aa-46d4-817a-f9c410238b45\") " pod="openstack/barbican-worker-756466fd6c-6vdmf" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.921519 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgjbc\" (UniqueName: \"kubernetes.io/projected/639973de-e4aa-46d4-817a-f9c410238b45-kube-api-access-zgjbc\") pod \"barbican-worker-756466fd6c-6vdmf\" (UID: \"639973de-e4aa-46d4-817a-f9c410238b45\") " pod="openstack/barbican-worker-756466fd6c-6vdmf" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.928902 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-6jw7g"] Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.965318 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/639973de-e4aa-46d4-817a-f9c410238b45-combined-ca-bundle\") pod \"barbican-worker-756466fd6c-6vdmf\" (UID: \"639973de-e4aa-46d4-817a-f9c410238b45\") " pod="openstack/barbican-worker-756466fd6c-6vdmf" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.965882 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/639973de-e4aa-46d4-817a-f9c410238b45-config-data-custom\") pod \"barbican-worker-756466fd6c-6vdmf\" (UID: \"639973de-e4aa-46d4-817a-f9c410238b45\") " pod="openstack/barbican-worker-756466fd6c-6vdmf" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.970746 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-6jw7g\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.970790 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-6jw7g\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.971363 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adeff71a-2b97-46e9-a945-d69cf06e34d9-logs\") pod \"barbican-keystone-listener-7c894bbd86-9phxx\" (UID: \"adeff71a-2b97-46e9-a945-d69cf06e34d9\") " pod="openstack/barbican-keystone-listener-7c894bbd86-9phxx" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.971416 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/adeff71a-2b97-46e9-a945-d69cf06e34d9-config-data-custom\") pod \"barbican-keystone-listener-7c894bbd86-9phxx\" (UID: \"adeff71a-2b97-46e9-a945-d69cf06e34d9\") " pod="openstack/barbican-keystone-listener-7c894bbd86-9phxx" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.971451 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adeff71a-2b97-46e9-a945-d69cf06e34d9-config-data\") pod \"barbican-keystone-listener-7c894bbd86-9phxx\" (UID: \"adeff71a-2b97-46e9-a945-d69cf06e34d9\") " pod="openstack/barbican-keystone-listener-7c894bbd86-9phxx" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.971479 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjjrc\" (UniqueName: \"kubernetes.io/projected/adeff71a-2b97-46e9-a945-d69cf06e34d9-kube-api-access-zjjrc\") pod \"barbican-keystone-listener-7c894bbd86-9phxx\" (UID: \"adeff71a-2b97-46e9-a945-d69cf06e34d9\") " pod="openstack/barbican-keystone-listener-7c894bbd86-9phxx" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.971553 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mmx2\" (UniqueName: \"kubernetes.io/projected/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-kube-api-access-2mmx2\") pod \"dnsmasq-dns-85ff748b95-6jw7g\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.971617 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-6jw7g\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.971651 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adeff71a-2b97-46e9-a945-d69cf06e34d9-combined-ca-bundle\") pod \"barbican-keystone-listener-7c894bbd86-9phxx\" (UID: \"adeff71a-2b97-46e9-a945-d69cf06e34d9\") " pod="openstack/barbican-keystone-listener-7c894bbd86-9phxx" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.971672 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-config\") pod \"dnsmasq-dns-85ff748b95-6jw7g\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.971749 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-dns-svc\") pod \"dnsmasq-dns-85ff748b95-6jw7g\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.972122 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/639973de-e4aa-46d4-817a-f9c410238b45-config-data\") pod \"barbican-worker-756466fd6c-6vdmf\" (UID: \"639973de-e4aa-46d4-817a-f9c410238b45\") " pod="openstack/barbican-worker-756466fd6c-6vdmf" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.972205 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adeff71a-2b97-46e9-a945-d69cf06e34d9-logs\") pod \"barbican-keystone-listener-7c894bbd86-9phxx\" (UID: \"adeff71a-2b97-46e9-a945-d69cf06e34d9\") " pod="openstack/barbican-keystone-listener-7c894bbd86-9phxx" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.988273 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adeff71a-2b97-46e9-a945-d69cf06e34d9-combined-ca-bundle\") pod \"barbican-keystone-listener-7c894bbd86-9phxx\" (UID: \"adeff71a-2b97-46e9-a945-d69cf06e34d9\") " pod="openstack/barbican-keystone-listener-7c894bbd86-9phxx" Mar 21 08:47:48 crc kubenswrapper[4696]: I0321 08:47:48.995967 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adeff71a-2b97-46e9-a945-d69cf06e34d9-config-data\") pod \"barbican-keystone-listener-7c894bbd86-9phxx\" (UID: \"adeff71a-2b97-46e9-a945-d69cf06e34d9\") " pod="openstack/barbican-keystone-listener-7c894bbd86-9phxx" Mar 21 08:47:48 crc kubenswrapper[4696]: W0321 08:47:48.997987 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31091b53_3bf3_4704_96bc_c5c33eb55218.slice/crio-90b49a87f5d8cb5e2acdc5e54e3ca26cf3700d15a5c8e535f8562f848bbc1002 WatchSource:0}: Error finding container 90b49a87f5d8cb5e2acdc5e54e3ca26cf3700d15a5c8e535f8562f848bbc1002: Status 404 returned error can't find the container with id 90b49a87f5d8cb5e2acdc5e54e3ca26cf3700d15a5c8e535f8562f848bbc1002 Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.000407 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/adeff71a-2b97-46e9-a945-d69cf06e34d9-config-data-custom\") pod \"barbican-keystone-listener-7c894bbd86-9phxx\" (UID: \"adeff71a-2b97-46e9-a945-d69cf06e34d9\") " pod="openstack/barbican-keystone-listener-7c894bbd86-9phxx" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.028616 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjjrc\" (UniqueName: \"kubernetes.io/projected/adeff71a-2b97-46e9-a945-d69cf06e34d9-kube-api-access-zjjrc\") pod \"barbican-keystone-listener-7c894bbd86-9phxx\" (UID: \"adeff71a-2b97-46e9-a945-d69cf06e34d9\") " pod="openstack/barbican-keystone-listener-7c894bbd86-9phxx" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.057155 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6b5744ff58-x5lj7"] Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.060902 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.069096 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.074031 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mmx2\" (UniqueName: \"kubernetes.io/projected/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-kube-api-access-2mmx2\") pod \"dnsmasq-dns-85ff748b95-6jw7g\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.074089 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-6jw7g\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.074116 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-config\") pod \"dnsmasq-dns-85ff748b95-6jw7g\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.074168 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-dns-svc\") pod \"dnsmasq-dns-85ff748b95-6jw7g\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.074214 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-6jw7g\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.074234 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-6jw7g\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.075078 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-6jw7g\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.075800 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-6jw7g\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.076325 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-config\") pod \"dnsmasq-dns-85ff748b95-6jw7g\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.077551 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6b5744ff58-x5lj7"] Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.077642 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-6jw7g\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.083308 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-dns-svc\") pod \"dnsmasq-dns-85ff748b95-6jw7g\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.132510 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mmx2\" (UniqueName: \"kubernetes.io/projected/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-kube-api-access-2mmx2\") pod \"dnsmasq-dns-85ff748b95-6jw7g\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.178115 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/876c7f89-b54b-4bcc-8220-33df4b204390-logs\") pod \"barbican-api-6b5744ff58-x5lj7\" (UID: \"876c7f89-b54b-4bcc-8220-33df4b204390\") " pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.178177 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxzvr\" (UniqueName: \"kubernetes.io/projected/876c7f89-b54b-4bcc-8220-33df4b204390-kube-api-access-dxzvr\") pod \"barbican-api-6b5744ff58-x5lj7\" (UID: \"876c7f89-b54b-4bcc-8220-33df4b204390\") " pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.178231 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/876c7f89-b54b-4bcc-8220-33df4b204390-combined-ca-bundle\") pod \"barbican-api-6b5744ff58-x5lj7\" (UID: \"876c7f89-b54b-4bcc-8220-33df4b204390\") " pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.178269 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/876c7f89-b54b-4bcc-8220-33df4b204390-config-data-custom\") pod \"barbican-api-6b5744ff58-x5lj7\" (UID: \"876c7f89-b54b-4bcc-8220-33df4b204390\") " pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.178313 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/876c7f89-b54b-4bcc-8220-33df4b204390-config-data\") pod \"barbican-api-6b5744ff58-x5lj7\" (UID: \"876c7f89-b54b-4bcc-8220-33df4b204390\") " pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.182988 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1af36c17-5501-4d4e-87a9-e9b43cb9eef0","Type":"ContainerStarted","Data":"c3f089a13d2570c4fe4cafb5efecab2fdcd188b1e190e48ec3fb875343fe9ab2"} Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.184878 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-696f547b5d-r9gcd" event={"ID":"31091b53-3bf3-4704-96bc-c5c33eb55218","Type":"ContainerStarted","Data":"90b49a87f5d8cb5e2acdc5e54e3ca26cf3700d15a5c8e535f8562f848bbc1002"} Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.184936 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.211623 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-756466fd6c-6vdmf" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.217136 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.281422 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/876c7f89-b54b-4bcc-8220-33df4b204390-config-data\") pod \"barbican-api-6b5744ff58-x5lj7\" (UID: \"876c7f89-b54b-4bcc-8220-33df4b204390\") " pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.281532 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/876c7f89-b54b-4bcc-8220-33df4b204390-logs\") pod \"barbican-api-6b5744ff58-x5lj7\" (UID: \"876c7f89-b54b-4bcc-8220-33df4b204390\") " pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.281574 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxzvr\" (UniqueName: \"kubernetes.io/projected/876c7f89-b54b-4bcc-8220-33df4b204390-kube-api-access-dxzvr\") pod \"barbican-api-6b5744ff58-x5lj7\" (UID: \"876c7f89-b54b-4bcc-8220-33df4b204390\") " pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.281621 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/876c7f89-b54b-4bcc-8220-33df4b204390-combined-ca-bundle\") pod \"barbican-api-6b5744ff58-x5lj7\" (UID: \"876c7f89-b54b-4bcc-8220-33df4b204390\") " pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.281654 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/876c7f89-b54b-4bcc-8220-33df4b204390-config-data-custom\") pod \"barbican-api-6b5744ff58-x5lj7\" (UID: \"876c7f89-b54b-4bcc-8220-33df4b204390\") " pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.283000 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/876c7f89-b54b-4bcc-8220-33df4b204390-logs\") pod \"barbican-api-6b5744ff58-x5lj7\" (UID: \"876c7f89-b54b-4bcc-8220-33df4b204390\") " pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.290078 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7c894bbd86-9phxx" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.307120 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/876c7f89-b54b-4bcc-8220-33df4b204390-config-data\") pod \"barbican-api-6b5744ff58-x5lj7\" (UID: \"876c7f89-b54b-4bcc-8220-33df4b204390\") " pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.308477 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/876c7f89-b54b-4bcc-8220-33df4b204390-config-data-custom\") pod \"barbican-api-6b5744ff58-x5lj7\" (UID: \"876c7f89-b54b-4bcc-8220-33df4b204390\") " pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.308510 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/876c7f89-b54b-4bcc-8220-33df4b204390-combined-ca-bundle\") pod \"barbican-api-6b5744ff58-x5lj7\" (UID: \"876c7f89-b54b-4bcc-8220-33df4b204390\") " pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.324659 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxzvr\" (UniqueName: \"kubernetes.io/projected/876c7f89-b54b-4bcc-8220-33df4b204390-kube-api-access-dxzvr\") pod \"barbican-api-6b5744ff58-x5lj7\" (UID: \"876c7f89-b54b-4bcc-8220-33df4b204390\") " pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.336543 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "64b14a78-860a-48c3-ae69-16cc36c77be8" (UID: "64b14a78-860a-48c3-ae69-16cc36c77be8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.358357 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "64b14a78-860a-48c3-ae69-16cc36c77be8" (UID: "64b14a78-860a-48c3-ae69-16cc36c77be8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.367245 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "64b14a78-860a-48c3-ae69-16cc36c77be8" (UID: "64b14a78-860a-48c3-ae69-16cc36c77be8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.383125 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.383156 4696 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.383168 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.385332 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-config" (OuterVolumeSpecName: "config") pod "64b14a78-860a-48c3-ae69-16cc36c77be8" (UID: "64b14a78-860a-48c3-ae69-16cc36c77be8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.488164 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64b14a78-860a-48c3-ae69-16cc36c77be8-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.586577 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.598388 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-mddgz"] Mar 21 08:47:49 crc kubenswrapper[4696]: I0321 08:47:49.640290 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-mddgz"] Mar 21 08:47:50 crc kubenswrapper[4696]: I0321 08:47:50.049018 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-6jw7g"] Mar 21 08:47:50 crc kubenswrapper[4696]: I0321 08:47:50.200919 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7c894bbd86-9phxx"] Mar 21 08:47:50 crc kubenswrapper[4696]: I0321 08:47:50.207878 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-756466fd6c-6vdmf"] Mar 21 08:47:50 crc kubenswrapper[4696]: I0321 08:47:50.209394 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-565c47bc8-bmdxv" event={"ID":"a32ac807-7955-425a-a971-e07ddde36bb0","Type":"ContainerStarted","Data":"753467b400aac31c890d0990cf0105c1387d9e9a97ee9e201734937908aa8474"} Mar 21 08:47:50 crc kubenswrapper[4696]: I0321 08:47:50.210913 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:50 crc kubenswrapper[4696]: I0321 08:47:50.210961 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:50 crc kubenswrapper[4696]: I0321 08:47:50.218678 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-696f547b5d-r9gcd" event={"ID":"31091b53-3bf3-4704-96bc-c5c33eb55218","Type":"ContainerStarted","Data":"1659fda4af2cb9786934a745e36386744e668574bf00f39f49b31e830cb69b30"} Mar 21 08:47:50 crc kubenswrapper[4696]: I0321 08:47:50.219050 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:47:50 crc kubenswrapper[4696]: I0321 08:47:50.230724 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-565c47bc8-bmdxv" podStartSLOduration=4.23070639 podStartE2EDuration="4.23070639s" podCreationTimestamp="2026-03-21 08:47:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:47:50.230527185 +0000 UTC m=+1204.351407918" watchObservedRunningTime="2026-03-21 08:47:50.23070639 +0000 UTC m=+1204.351587103" Mar 21 08:47:50 crc kubenswrapper[4696]: I0321 08:47:50.263515 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-696f547b5d-r9gcd" podStartSLOduration=3.263494894 podStartE2EDuration="3.263494894s" podCreationTimestamp="2026-03-21 08:47:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:47:50.249620857 +0000 UTC m=+1204.370501570" watchObservedRunningTime="2026-03-21 08:47:50.263494894 +0000 UTC m=+1204.384375607" Mar 21 08:47:50 crc kubenswrapper[4696]: I0321 08:47:50.453461 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6b5744ff58-x5lj7"] Mar 21 08:47:50 crc kubenswrapper[4696]: I0321 08:47:50.548210 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64b14a78-860a-48c3-ae69-16cc36c77be8" path="/var/lib/kubelet/pods/64b14a78-860a-48c3-ae69-16cc36c77be8/volumes" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.229509 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1af36c17-5501-4d4e-87a9-e9b43cb9eef0","Type":"ContainerStarted","Data":"530a9760edef3bd29be8445624f8167d7d983c70e52b38653480e80e7a5c7b3a"} Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.251627 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.251611738 podStartE2EDuration="6.251611738s" podCreationTimestamp="2026-03-21 08:47:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:47:51.249662684 +0000 UTC m=+1205.370543397" watchObservedRunningTime="2026-03-21 08:47:51.251611738 +0000 UTC m=+1205.372492451" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.430987 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7dd88fbdb6-nnxhg"] Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.437321 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.461775 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.462131 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.498515 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7dd88fbdb6-nnxhg"] Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.554401 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd0e5187-2172-48fc-90e8-4ce8f3f1c143-config-data-custom\") pod \"barbican-api-7dd88fbdb6-nnxhg\" (UID: \"bd0e5187-2172-48fc-90e8-4ce8f3f1c143\") " pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.554487 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd0e5187-2172-48fc-90e8-4ce8f3f1c143-combined-ca-bundle\") pod \"barbican-api-7dd88fbdb6-nnxhg\" (UID: \"bd0e5187-2172-48fc-90e8-4ce8f3f1c143\") " pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.554515 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd0e5187-2172-48fc-90e8-4ce8f3f1c143-public-tls-certs\") pod \"barbican-api-7dd88fbdb6-nnxhg\" (UID: \"bd0e5187-2172-48fc-90e8-4ce8f3f1c143\") " pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.554568 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd0e5187-2172-48fc-90e8-4ce8f3f1c143-logs\") pod \"barbican-api-7dd88fbdb6-nnxhg\" (UID: \"bd0e5187-2172-48fc-90e8-4ce8f3f1c143\") " pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.554588 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd0e5187-2172-48fc-90e8-4ce8f3f1c143-internal-tls-certs\") pod \"barbican-api-7dd88fbdb6-nnxhg\" (UID: \"bd0e5187-2172-48fc-90e8-4ce8f3f1c143\") " pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.554624 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x2q8\" (UniqueName: \"kubernetes.io/projected/bd0e5187-2172-48fc-90e8-4ce8f3f1c143-kube-api-access-6x2q8\") pod \"barbican-api-7dd88fbdb6-nnxhg\" (UID: \"bd0e5187-2172-48fc-90e8-4ce8f3f1c143\") " pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.554666 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd0e5187-2172-48fc-90e8-4ce8f3f1c143-config-data\") pod \"barbican-api-7dd88fbdb6-nnxhg\" (UID: \"bd0e5187-2172-48fc-90e8-4ce8f3f1c143\") " pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.657255 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd0e5187-2172-48fc-90e8-4ce8f3f1c143-combined-ca-bundle\") pod \"barbican-api-7dd88fbdb6-nnxhg\" (UID: \"bd0e5187-2172-48fc-90e8-4ce8f3f1c143\") " pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.657337 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd0e5187-2172-48fc-90e8-4ce8f3f1c143-public-tls-certs\") pod \"barbican-api-7dd88fbdb6-nnxhg\" (UID: \"bd0e5187-2172-48fc-90e8-4ce8f3f1c143\") " pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.657496 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd0e5187-2172-48fc-90e8-4ce8f3f1c143-logs\") pod \"barbican-api-7dd88fbdb6-nnxhg\" (UID: \"bd0e5187-2172-48fc-90e8-4ce8f3f1c143\") " pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.657524 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd0e5187-2172-48fc-90e8-4ce8f3f1c143-internal-tls-certs\") pod \"barbican-api-7dd88fbdb6-nnxhg\" (UID: \"bd0e5187-2172-48fc-90e8-4ce8f3f1c143\") " pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.657606 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x2q8\" (UniqueName: \"kubernetes.io/projected/bd0e5187-2172-48fc-90e8-4ce8f3f1c143-kube-api-access-6x2q8\") pod \"barbican-api-7dd88fbdb6-nnxhg\" (UID: \"bd0e5187-2172-48fc-90e8-4ce8f3f1c143\") " pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.657699 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd0e5187-2172-48fc-90e8-4ce8f3f1c143-config-data\") pod \"barbican-api-7dd88fbdb6-nnxhg\" (UID: \"bd0e5187-2172-48fc-90e8-4ce8f3f1c143\") " pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.657774 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd0e5187-2172-48fc-90e8-4ce8f3f1c143-config-data-custom\") pod \"barbican-api-7dd88fbdb6-nnxhg\" (UID: \"bd0e5187-2172-48fc-90e8-4ce8f3f1c143\") " pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.658310 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd0e5187-2172-48fc-90e8-4ce8f3f1c143-logs\") pod \"barbican-api-7dd88fbdb6-nnxhg\" (UID: \"bd0e5187-2172-48fc-90e8-4ce8f3f1c143\") " pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.669861 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd0e5187-2172-48fc-90e8-4ce8f3f1c143-combined-ca-bundle\") pod \"barbican-api-7dd88fbdb6-nnxhg\" (UID: \"bd0e5187-2172-48fc-90e8-4ce8f3f1c143\") " pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.674723 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd0e5187-2172-48fc-90e8-4ce8f3f1c143-internal-tls-certs\") pod \"barbican-api-7dd88fbdb6-nnxhg\" (UID: \"bd0e5187-2172-48fc-90e8-4ce8f3f1c143\") " pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.709671 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd0e5187-2172-48fc-90e8-4ce8f3f1c143-public-tls-certs\") pod \"barbican-api-7dd88fbdb6-nnxhg\" (UID: \"bd0e5187-2172-48fc-90e8-4ce8f3f1c143\") " pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.711334 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd0e5187-2172-48fc-90e8-4ce8f3f1c143-config-data\") pod \"barbican-api-7dd88fbdb6-nnxhg\" (UID: \"bd0e5187-2172-48fc-90e8-4ce8f3f1c143\") " pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.715116 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd0e5187-2172-48fc-90e8-4ce8f3f1c143-config-data-custom\") pod \"barbican-api-7dd88fbdb6-nnxhg\" (UID: \"bd0e5187-2172-48fc-90e8-4ce8f3f1c143\") " pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.719236 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x2q8\" (UniqueName: \"kubernetes.io/projected/bd0e5187-2172-48fc-90e8-4ce8f3f1c143-kube-api-access-6x2q8\") pod \"barbican-api-7dd88fbdb6-nnxhg\" (UID: \"bd0e5187-2172-48fc-90e8-4ce8f3f1c143\") " pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:51 crc kubenswrapper[4696]: I0321 08:47:51.784536 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:53 crc kubenswrapper[4696]: I0321 08:47:53.043492 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-mddgz" podUID="64b14a78-860a-48c3-ae69-16cc36c77be8" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.167:5353: i/o timeout" Mar 21 08:47:54 crc kubenswrapper[4696]: I0321 08:47:54.061288 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:54 crc kubenswrapper[4696]: I0321 08:47:54.269154 4696 generic.go:334] "Generic (PLEG): container finished" podID="d3fc5362-53e5-4c35-be23-49dc3dd1fcba" containerID="83e39ef69714631ade43ac7b3e551d20a5d282e9d803ae8961e6ae213c1e452e" exitCode=0 Mar 21 08:47:54 crc kubenswrapper[4696]: I0321 08:47:54.269222 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-vgsz6" event={"ID":"d3fc5362-53e5-4c35-be23-49dc3dd1fcba","Type":"ContainerDied","Data":"83e39ef69714631ade43ac7b3e551d20a5d282e9d803ae8961e6ae213c1e452e"} Mar 21 08:47:54 crc kubenswrapper[4696]: W0321 08:47:54.317059 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0e814b1_d0f8_4a97_a76e_da97f375c9a9.slice/crio-5153db0aeb397e48b4ed0cffa78cfe6af26091044bf749685d6054821e78a5d9 WatchSource:0}: Error finding container 5153db0aeb397e48b4ed0cffa78cfe6af26091044bf749685d6054821e78a5d9: Status 404 returned error can't find the container with id 5153db0aeb397e48b4ed0cffa78cfe6af26091044bf749685d6054821e78a5d9 Mar 21 08:47:54 crc kubenswrapper[4696]: W0321 08:47:54.330084 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod639973de_e4aa_46d4_817a_f9c410238b45.slice/crio-cbbdaaa8800996e31c20e8b0f721687f905953fc0576a74a142180f7f5dd729c WatchSource:0}: Error finding container cbbdaaa8800996e31c20e8b0f721687f905953fc0576a74a142180f7f5dd729c: Status 404 returned error can't find the container with id cbbdaaa8800996e31c20e8b0f721687f905953fc0576a74a142180f7f5dd729c Mar 21 08:47:54 crc kubenswrapper[4696]: W0321 08:47:54.338988 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod876c7f89_b54b_4bcc_8220_33df4b204390.slice/crio-ee9d5ed650ca674ed29b53bc011acaf13cc16e9e5f48e0a71f58c8ad2a67cadc WatchSource:0}: Error finding container ee9d5ed650ca674ed29b53bc011acaf13cc16e9e5f48e0a71f58c8ad2a67cadc: Status 404 returned error can't find the container with id ee9d5ed650ca674ed29b53bc011acaf13cc16e9e5f48e0a71f58c8ad2a67cadc Mar 21 08:47:54 crc kubenswrapper[4696]: W0321 08:47:54.995726 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd0e5187_2172_48fc_90e8_4ce8f3f1c143.slice/crio-578722825be40f73b932a85f298a671064c0822142831ef332efbf7cfb9c0ee5 WatchSource:0}: Error finding container 578722825be40f73b932a85f298a671064c0822142831ef332efbf7cfb9c0ee5: Status 404 returned error can't find the container with id 578722825be40f73b932a85f298a671064c0822142831ef332efbf7cfb9c0ee5 Mar 21 08:47:55 crc kubenswrapper[4696]: I0321 08:47:55.002600 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7dd88fbdb6-nnxhg"] Mar 21 08:47:55 crc kubenswrapper[4696]: I0321 08:47:55.296845 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbd625f3-5fcc-48f5-9022-df08fdc01887","Type":"ContainerStarted","Data":"2e4eed46e1a4de0167e86432becd21c8cb3965e91eaa6b63e2084ca18ed3b13e"} Mar 21 08:47:55 crc kubenswrapper[4696]: I0321 08:47:55.306163 4696 generic.go:334] "Generic (PLEG): container finished" podID="c0e814b1-d0f8-4a97-a76e-da97f375c9a9" containerID="bf696af1ec291a1084ebc174abfdd15a2b74d5894459e8b56e1bc8835cac1e17" exitCode=0 Mar 21 08:47:55 crc kubenswrapper[4696]: I0321 08:47:55.306247 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" event={"ID":"c0e814b1-d0f8-4a97-a76e-da97f375c9a9","Type":"ContainerDied","Data":"bf696af1ec291a1084ebc174abfdd15a2b74d5894459e8b56e1bc8835cac1e17"} Mar 21 08:47:55 crc kubenswrapper[4696]: I0321 08:47:55.306273 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" event={"ID":"c0e814b1-d0f8-4a97-a76e-da97f375c9a9","Type":"ContainerStarted","Data":"5153db0aeb397e48b4ed0cffa78cfe6af26091044bf749685d6054821e78a5d9"} Mar 21 08:47:55 crc kubenswrapper[4696]: I0321 08:47:55.318073 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b5744ff58-x5lj7" event={"ID":"876c7f89-b54b-4bcc-8220-33df4b204390","Type":"ContainerStarted","Data":"5ea1280d4db7c529c5b86b33c1e58f42230d8eb4ccbf68af77facb55ec027f7f"} Mar 21 08:47:55 crc kubenswrapper[4696]: I0321 08:47:55.318118 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b5744ff58-x5lj7" event={"ID":"876c7f89-b54b-4bcc-8220-33df4b204390","Type":"ContainerStarted","Data":"ee9d5ed650ca674ed29b53bc011acaf13cc16e9e5f48e0a71f58c8ad2a67cadc"} Mar 21 08:47:55 crc kubenswrapper[4696]: I0321 08:47:55.365443 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7c894bbd86-9phxx" event={"ID":"adeff71a-2b97-46e9-a945-d69cf06e34d9","Type":"ContainerStarted","Data":"b50b0cd37802bcb2cf759eb4e281e2b58b53b1a996bf8008d59a1b2c3d65bc0f"} Mar 21 08:47:55 crc kubenswrapper[4696]: I0321 08:47:55.375765 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7dd88fbdb6-nnxhg" event={"ID":"bd0e5187-2172-48fc-90e8-4ce8f3f1c143","Type":"ContainerStarted","Data":"83d40800b4b10aa23360c80d23bec523ba23adcdae3afb7000f70ae8cbf10a72"} Mar 21 08:47:55 crc kubenswrapper[4696]: I0321 08:47:55.375809 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7dd88fbdb6-nnxhg" event={"ID":"bd0e5187-2172-48fc-90e8-4ce8f3f1c143","Type":"ContainerStarted","Data":"578722825be40f73b932a85f298a671064c0822142831ef332efbf7cfb9c0ee5"} Mar 21 08:47:55 crc kubenswrapper[4696]: I0321 08:47:55.378388 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-756466fd6c-6vdmf" event={"ID":"639973de-e4aa-46d4-817a-f9c410238b45","Type":"ContainerStarted","Data":"cbbdaaa8800996e31c20e8b0f721687f905953fc0576a74a142180f7f5dd729c"} Mar 21 08:47:55 crc kubenswrapper[4696]: I0321 08:47:55.583294 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:47:55 crc kubenswrapper[4696]: I0321 08:47:55.794087 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:55 crc kubenswrapper[4696]: I0321 08:47:55.794128 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:55 crc kubenswrapper[4696]: I0321 08:47:55.838356 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:55 crc kubenswrapper[4696]: I0321 08:47:55.856095 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:55 crc kubenswrapper[4696]: I0321 08:47:55.961878 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-56b7788dc4-f928g"] Mar 21 08:47:55 crc kubenswrapper[4696]: I0321 08:47:55.963592 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:55 crc kubenswrapper[4696]: I0321 08:47:55.991177 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-56b7788dc4-f928g"] Mar 21 08:47:55 crc kubenswrapper[4696]: I0321 08:47:55.992242 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-vgsz6" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.066924 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-config-data\") pod \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\" (UID: \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\") " Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.066994 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-scripts\") pod \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\" (UID: \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\") " Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.067104 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-certs\") pod \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\" (UID: \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\") " Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.068022 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-combined-ca-bundle\") pod \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\" (UID: \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\") " Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.068710 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mv7dw\" (UniqueName: \"kubernetes.io/projected/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-kube-api-access-mv7dw\") pod \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\" (UID: \"d3fc5362-53e5-4c35-be23-49dc3dd1fcba\") " Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.069022 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bd18ca4-e084-4fe1-85e3-7319ddd703ff-scripts\") pod \"placement-56b7788dc4-f928g\" (UID: \"3bd18ca4-e084-4fe1-85e3-7319ddd703ff\") " pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.070435 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bd18ca4-e084-4fe1-85e3-7319ddd703ff-public-tls-certs\") pod \"placement-56b7788dc4-f928g\" (UID: \"3bd18ca4-e084-4fe1-85e3-7319ddd703ff\") " pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.070552 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bd18ca4-e084-4fe1-85e3-7319ddd703ff-logs\") pod \"placement-56b7788dc4-f928g\" (UID: \"3bd18ca4-e084-4fe1-85e3-7319ddd703ff\") " pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.070617 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bd18ca4-e084-4fe1-85e3-7319ddd703ff-internal-tls-certs\") pod \"placement-56b7788dc4-f928g\" (UID: \"3bd18ca4-e084-4fe1-85e3-7319ddd703ff\") " pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.070665 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bd18ca4-e084-4fe1-85e3-7319ddd703ff-combined-ca-bundle\") pod \"placement-56b7788dc4-f928g\" (UID: \"3bd18ca4-e084-4fe1-85e3-7319ddd703ff\") " pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.070682 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgxdm\" (UniqueName: \"kubernetes.io/projected/3bd18ca4-e084-4fe1-85e3-7319ddd703ff-kube-api-access-pgxdm\") pod \"placement-56b7788dc4-f928g\" (UID: \"3bd18ca4-e084-4fe1-85e3-7319ddd703ff\") " pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.070730 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bd18ca4-e084-4fe1-85e3-7319ddd703ff-config-data\") pod \"placement-56b7788dc4-f928g\" (UID: \"3bd18ca4-e084-4fe1-85e3-7319ddd703ff\") " pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.092561 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-kube-api-access-mv7dw" (OuterVolumeSpecName: "kube-api-access-mv7dw") pod "d3fc5362-53e5-4c35-be23-49dc3dd1fcba" (UID: "d3fc5362-53e5-4c35-be23-49dc3dd1fcba"). InnerVolumeSpecName "kube-api-access-mv7dw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.093380 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-certs" (OuterVolumeSpecName: "certs") pod "d3fc5362-53e5-4c35-be23-49dc3dd1fcba" (UID: "d3fc5362-53e5-4c35-be23-49dc3dd1fcba"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.110285 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3fc5362-53e5-4c35-be23-49dc3dd1fcba" (UID: "d3fc5362-53e5-4c35-be23-49dc3dd1fcba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.118999 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-scripts" (OuterVolumeSpecName: "scripts") pod "d3fc5362-53e5-4c35-be23-49dc3dd1fcba" (UID: "d3fc5362-53e5-4c35-be23-49dc3dd1fcba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.141671 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-config-data" (OuterVolumeSpecName: "config-data") pod "d3fc5362-53e5-4c35-be23-49dc3dd1fcba" (UID: "d3fc5362-53e5-4c35-be23-49dc3dd1fcba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.174947 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bd18ca4-e084-4fe1-85e3-7319ddd703ff-public-tls-certs\") pod \"placement-56b7788dc4-f928g\" (UID: \"3bd18ca4-e084-4fe1-85e3-7319ddd703ff\") " pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.175049 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bd18ca4-e084-4fe1-85e3-7319ddd703ff-logs\") pod \"placement-56b7788dc4-f928g\" (UID: \"3bd18ca4-e084-4fe1-85e3-7319ddd703ff\") " pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.175096 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bd18ca4-e084-4fe1-85e3-7319ddd703ff-internal-tls-certs\") pod \"placement-56b7788dc4-f928g\" (UID: \"3bd18ca4-e084-4fe1-85e3-7319ddd703ff\") " pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.175135 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bd18ca4-e084-4fe1-85e3-7319ddd703ff-combined-ca-bundle\") pod \"placement-56b7788dc4-f928g\" (UID: \"3bd18ca4-e084-4fe1-85e3-7319ddd703ff\") " pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.175156 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgxdm\" (UniqueName: \"kubernetes.io/projected/3bd18ca4-e084-4fe1-85e3-7319ddd703ff-kube-api-access-pgxdm\") pod \"placement-56b7788dc4-f928g\" (UID: \"3bd18ca4-e084-4fe1-85e3-7319ddd703ff\") " pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.175182 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bd18ca4-e084-4fe1-85e3-7319ddd703ff-config-data\") pod \"placement-56b7788dc4-f928g\" (UID: \"3bd18ca4-e084-4fe1-85e3-7319ddd703ff\") " pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.175227 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bd18ca4-e084-4fe1-85e3-7319ddd703ff-scripts\") pod \"placement-56b7788dc4-f928g\" (UID: \"3bd18ca4-e084-4fe1-85e3-7319ddd703ff\") " pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.175350 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.175370 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.175384 4696 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.175396 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.175408 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mv7dw\" (UniqueName: \"kubernetes.io/projected/d3fc5362-53e5-4c35-be23-49dc3dd1fcba-kube-api-access-mv7dw\") on node \"crc\" DevicePath \"\"" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.182355 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bd18ca4-e084-4fe1-85e3-7319ddd703ff-scripts\") pod \"placement-56b7788dc4-f928g\" (UID: \"3bd18ca4-e084-4fe1-85e3-7319ddd703ff\") " pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.187134 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bd18ca4-e084-4fe1-85e3-7319ddd703ff-internal-tls-certs\") pod \"placement-56b7788dc4-f928g\" (UID: \"3bd18ca4-e084-4fe1-85e3-7319ddd703ff\") " pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.187426 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bd18ca4-e084-4fe1-85e3-7319ddd703ff-combined-ca-bundle\") pod \"placement-56b7788dc4-f928g\" (UID: \"3bd18ca4-e084-4fe1-85e3-7319ddd703ff\") " pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.187475 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bd18ca4-e084-4fe1-85e3-7319ddd703ff-logs\") pod \"placement-56b7788dc4-f928g\" (UID: \"3bd18ca4-e084-4fe1-85e3-7319ddd703ff\") " pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.191415 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bd18ca4-e084-4fe1-85e3-7319ddd703ff-config-data\") pod \"placement-56b7788dc4-f928g\" (UID: \"3bd18ca4-e084-4fe1-85e3-7319ddd703ff\") " pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.193712 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bd18ca4-e084-4fe1-85e3-7319ddd703ff-public-tls-certs\") pod \"placement-56b7788dc4-f928g\" (UID: \"3bd18ca4-e084-4fe1-85e3-7319ddd703ff\") " pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.210393 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgxdm\" (UniqueName: \"kubernetes.io/projected/3bd18ca4-e084-4fe1-85e3-7319ddd703ff-kube-api-access-pgxdm\") pod \"placement-56b7788dc4-f928g\" (UID: \"3bd18ca4-e084-4fe1-85e3-7319ddd703ff\") " pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.303162 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.393254 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-n5tdf" event={"ID":"a70d1c8a-2db3-4ade-af61-fc8225262c18","Type":"ContainerStarted","Data":"499e153025457892a2d3dd1ae7d23b19b1557d52982f153cdf92e437caa1cd13"} Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.397425 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-vgsz6" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.397432 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-vgsz6" event={"ID":"d3fc5362-53e5-4c35-be23-49dc3dd1fcba","Type":"ContainerDied","Data":"bbfd9c99355e5907665a8e207c943040f2efbaba94ae5f6bf00a922e2aad9fdd"} Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.397471 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbfd9c99355e5907665a8e207c943040f2efbaba94ae5f6bf00a922e2aad9fdd" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.401187 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b5744ff58-x5lj7" event={"ID":"876c7f89-b54b-4bcc-8220-33df4b204390","Type":"ContainerStarted","Data":"be0507642215e219f86a3c60e4e54ca5611563c572f345e8fd5b997b4ed7b90d"} Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.401803 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.401884 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.404522 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" event={"ID":"c0e814b1-d0f8-4a97-a76e-da97f375c9a9","Type":"ContainerStarted","Data":"d7e8f9fab679381fe54039871691160d657f02791447d84979fe4b5b47767dc0"} Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.405268 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.416686 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7dd88fbdb6-nnxhg" event={"ID":"bd0e5187-2172-48fc-90e8-4ce8f3f1c143","Type":"ContainerStarted","Data":"6a978e289fab8ab59a22f810b2158a0c05f2a40374997be70c1c4d7889b77a6b"} Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.417114 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.417152 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.424789 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-n5tdf" podStartSLOduration=3.86880235 podStartE2EDuration="54.424770048s" podCreationTimestamp="2026-03-21 08:47:02 +0000 UTC" firstStartedPulling="2026-03-21 08:47:03.986659777 +0000 UTC m=+1158.107540490" lastFinishedPulling="2026-03-21 08:47:54.542627475 +0000 UTC m=+1208.663508188" observedRunningTime="2026-03-21 08:47:56.409997525 +0000 UTC m=+1210.530878248" watchObservedRunningTime="2026-03-21 08:47:56.424770048 +0000 UTC m=+1210.545650761" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.450925 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6b5744ff58-x5lj7" podStartSLOduration=8.450903937 podStartE2EDuration="8.450903937s" podCreationTimestamp="2026-03-21 08:47:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:47:56.431806964 +0000 UTC m=+1210.552687677" watchObservedRunningTime="2026-03-21 08:47:56.450903937 +0000 UTC m=+1210.571784650" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.504653 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" podStartSLOduration=8.504635427 podStartE2EDuration="8.504635427s" podCreationTimestamp="2026-03-21 08:47:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:47:56.464347032 +0000 UTC m=+1210.585227745" watchObservedRunningTime="2026-03-21 08:47:56.504635427 +0000 UTC m=+1210.625516140" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.526394 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-storageinit-8r5xw"] Mar 21 08:47:56 crc kubenswrapper[4696]: E0321 08:47:56.527158 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3fc5362-53e5-4c35-be23-49dc3dd1fcba" containerName="cloudkitty-db-sync" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.527271 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3fc5362-53e5-4c35-be23-49dc3dd1fcba" containerName="cloudkitty-db-sync" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.527523 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3fc5362-53e5-4c35-be23-49dc3dd1fcba" containerName="cloudkitty-db-sync" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.528523 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-8r5xw" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.575603 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.575912 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.576049 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.576152 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-ff6v6" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.577120 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.618903 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d73897de-7ac5-4991-aabf-b36cc6ef424d-combined-ca-bundle\") pod \"cloudkitty-storageinit-8r5xw\" (UID: \"d73897de-7ac5-4991-aabf-b36cc6ef424d\") " pod="openstack/cloudkitty-storageinit-8r5xw" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.619186 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d73897de-7ac5-4991-aabf-b36cc6ef424d-config-data\") pod \"cloudkitty-storageinit-8r5xw\" (UID: \"d73897de-7ac5-4991-aabf-b36cc6ef424d\") " pod="openstack/cloudkitty-storageinit-8r5xw" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.619232 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d73897de-7ac5-4991-aabf-b36cc6ef424d-scripts\") pod \"cloudkitty-storageinit-8r5xw\" (UID: \"d73897de-7ac5-4991-aabf-b36cc6ef424d\") " pod="openstack/cloudkitty-storageinit-8r5xw" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.619402 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d73897de-7ac5-4991-aabf-b36cc6ef424d-certs\") pod \"cloudkitty-storageinit-8r5xw\" (UID: \"d73897de-7ac5-4991-aabf-b36cc6ef424d\") " pod="openstack/cloudkitty-storageinit-8r5xw" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.619440 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lchv\" (UniqueName: \"kubernetes.io/projected/d73897de-7ac5-4991-aabf-b36cc6ef424d-kube-api-access-5lchv\") pod \"cloudkitty-storageinit-8r5xw\" (UID: \"d73897de-7ac5-4991-aabf-b36cc6ef424d\") " pod="openstack/cloudkitty-storageinit-8r5xw" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.633011 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7dd88fbdb6-nnxhg" podStartSLOduration=5.632991518 podStartE2EDuration="5.632991518s" podCreationTimestamp="2026-03-21 08:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:47:56.577613073 +0000 UTC m=+1210.698493796" watchObservedRunningTime="2026-03-21 08:47:56.632991518 +0000 UTC m=+1210.753872231" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.665334 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-8r5xw"] Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.724265 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d73897de-7ac5-4991-aabf-b36cc6ef424d-config-data\") pod \"cloudkitty-storageinit-8r5xw\" (UID: \"d73897de-7ac5-4991-aabf-b36cc6ef424d\") " pod="openstack/cloudkitty-storageinit-8r5xw" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.724329 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d73897de-7ac5-4991-aabf-b36cc6ef424d-scripts\") pod \"cloudkitty-storageinit-8r5xw\" (UID: \"d73897de-7ac5-4991-aabf-b36cc6ef424d\") " pod="openstack/cloudkitty-storageinit-8r5xw" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.724419 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d73897de-7ac5-4991-aabf-b36cc6ef424d-certs\") pod \"cloudkitty-storageinit-8r5xw\" (UID: \"d73897de-7ac5-4991-aabf-b36cc6ef424d\") " pod="openstack/cloudkitty-storageinit-8r5xw" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.724445 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lchv\" (UniqueName: \"kubernetes.io/projected/d73897de-7ac5-4991-aabf-b36cc6ef424d-kube-api-access-5lchv\") pod \"cloudkitty-storageinit-8r5xw\" (UID: \"d73897de-7ac5-4991-aabf-b36cc6ef424d\") " pod="openstack/cloudkitty-storageinit-8r5xw" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.724488 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d73897de-7ac5-4991-aabf-b36cc6ef424d-combined-ca-bundle\") pod \"cloudkitty-storageinit-8r5xw\" (UID: \"d73897de-7ac5-4991-aabf-b36cc6ef424d\") " pod="openstack/cloudkitty-storageinit-8r5xw" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.740209 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d73897de-7ac5-4991-aabf-b36cc6ef424d-config-data\") pod \"cloudkitty-storageinit-8r5xw\" (UID: \"d73897de-7ac5-4991-aabf-b36cc6ef424d\") " pod="openstack/cloudkitty-storageinit-8r5xw" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.740462 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d73897de-7ac5-4991-aabf-b36cc6ef424d-combined-ca-bundle\") pod \"cloudkitty-storageinit-8r5xw\" (UID: \"d73897de-7ac5-4991-aabf-b36cc6ef424d\") " pod="openstack/cloudkitty-storageinit-8r5xw" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.747525 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d73897de-7ac5-4991-aabf-b36cc6ef424d-certs\") pod \"cloudkitty-storageinit-8r5xw\" (UID: \"d73897de-7ac5-4991-aabf-b36cc6ef424d\") " pod="openstack/cloudkitty-storageinit-8r5xw" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.757504 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d73897de-7ac5-4991-aabf-b36cc6ef424d-scripts\") pod \"cloudkitty-storageinit-8r5xw\" (UID: \"d73897de-7ac5-4991-aabf-b36cc6ef424d\") " pod="openstack/cloudkitty-storageinit-8r5xw" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.768310 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lchv\" (UniqueName: \"kubernetes.io/projected/d73897de-7ac5-4991-aabf-b36cc6ef424d-kube-api-access-5lchv\") pod \"cloudkitty-storageinit-8r5xw\" (UID: \"d73897de-7ac5-4991-aabf-b36cc6ef424d\") " pod="openstack/cloudkitty-storageinit-8r5xw" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.791769 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.791849 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.905922 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-8r5xw" Mar 21 08:47:56 crc kubenswrapper[4696]: I0321 08:47:56.908149 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-56b7788dc4-f928g"] Mar 21 08:47:57 crc kubenswrapper[4696]: I0321 08:47:57.438906 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-56b7788dc4-f928g" event={"ID":"3bd18ca4-e084-4fe1-85e3-7319ddd703ff","Type":"ContainerStarted","Data":"0f4024764e0a862c9b8311ca3a6bc8688d177c56d2e358bfc7e0c6586e6d9d21"} Mar 21 08:47:57 crc kubenswrapper[4696]: I0321 08:47:57.476941 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-8r5xw"] Mar 21 08:47:58 crc kubenswrapper[4696]: I0321 08:47:58.466263 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-56b7788dc4-f928g" event={"ID":"3bd18ca4-e084-4fe1-85e3-7319ddd703ff","Type":"ContainerStarted","Data":"537aea56dd76971c15021303b8559ade451816dde92d0f089b146d05a24a6f5f"} Mar 21 08:47:58 crc kubenswrapper[4696]: I0321 08:47:58.468750 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-8r5xw" event={"ID":"d73897de-7ac5-4991-aabf-b36cc6ef424d","Type":"ContainerStarted","Data":"d6a2c9083ae133bb5944b16dc225cccbf9e3a5f9ea56291a7f5dd48fc657a5ad"} Mar 21 08:47:58 crc kubenswrapper[4696]: I0321 08:47:58.468796 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-8r5xw" event={"ID":"d73897de-7ac5-4991-aabf-b36cc6ef424d","Type":"ContainerStarted","Data":"8f426eb0e0c6a64f2e9272acebaaa0ce580d65fda3623c1726c9f20f26a79fdb"} Mar 21 08:47:58 crc kubenswrapper[4696]: I0321 08:47:58.492063 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-storageinit-8r5xw" podStartSLOduration=2.492044057 podStartE2EDuration="2.492044057s" podCreationTimestamp="2026-03-21 08:47:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:47:58.49035776 +0000 UTC m=+1212.611238473" watchObservedRunningTime="2026-03-21 08:47:58.492044057 +0000 UTC m=+1212.612924770" Mar 21 08:47:58 crc kubenswrapper[4696]: I0321 08:47:58.954916 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 21 08:47:58 crc kubenswrapper[4696]: I0321 08:47:58.955238 4696 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 21 08:47:59 crc kubenswrapper[4696]: I0321 08:47:59.052200 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.141347 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568048-j276n"] Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.143371 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568048-j276n" Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.152868 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568048-j276n"] Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.154335 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.154398 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.154453 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.220351 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6vxm\" (UniqueName: \"kubernetes.io/projected/41361f63-e76c-42d2-ae18-32c5e772773a-kube-api-access-r6vxm\") pod \"auto-csr-approver-29568048-j276n\" (UID: \"41361f63-e76c-42d2-ae18-32c5e772773a\") " pod="openshift-infra/auto-csr-approver-29568048-j276n" Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.322650 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6vxm\" (UniqueName: \"kubernetes.io/projected/41361f63-e76c-42d2-ae18-32c5e772773a-kube-api-access-r6vxm\") pod \"auto-csr-approver-29568048-j276n\" (UID: \"41361f63-e76c-42d2-ae18-32c5e772773a\") " pod="openshift-infra/auto-csr-approver-29568048-j276n" Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.339458 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6vxm\" (UniqueName: \"kubernetes.io/projected/41361f63-e76c-42d2-ae18-32c5e772773a-kube-api-access-r6vxm\") pod \"auto-csr-approver-29568048-j276n\" (UID: \"41361f63-e76c-42d2-ae18-32c5e772773a\") " pod="openshift-infra/auto-csr-approver-29568048-j276n" Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.341304 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.341367 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.341409 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.342504 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"37dc28f551f9db6ffb052a7a69754f617535fa9f02cee7492942ac3ed542a742"} pod="openshift-machine-config-operator/machine-config-daemon-z7srw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.342571 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" containerID="cri-o://37dc28f551f9db6ffb052a7a69754f617535fa9f02cee7492942ac3ed542a742" gracePeriod=600 Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.464654 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568048-j276n" Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.492123 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7c894bbd86-9phxx" event={"ID":"adeff71a-2b97-46e9-a945-d69cf06e34d9","Type":"ContainerStarted","Data":"5e01c257d057a31879b48ce96e77bf11a3d438f1606b4d92d0387b60fe5e0f73"} Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.492167 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7c894bbd86-9phxx" event={"ID":"adeff71a-2b97-46e9-a945-d69cf06e34d9","Type":"ContainerStarted","Data":"911bee93142a37f624948c874e7e8358b1f0e851d456fe67ae17dfb3aabd63f2"} Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.496016 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-756466fd6c-6vdmf" event={"ID":"639973de-e4aa-46d4-817a-f9c410238b45","Type":"ContainerStarted","Data":"283119d9db01df47ae6b11263b4b9a93351a24998da08a8b4e9f8c8d6427542d"} Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.496043 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-756466fd6c-6vdmf" event={"ID":"639973de-e4aa-46d4-817a-f9c410238b45","Type":"ContainerStarted","Data":"bffeec240df7f072e44f5b18f9127c6d72e8ae2de60fe8730337b55d226ec324"} Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.499511 4696 generic.go:334] "Generic (PLEG): container finished" podID="daaf227a-2305-495e-8495-a280abcd8e10" containerID="37dc28f551f9db6ffb052a7a69754f617535fa9f02cee7492942ac3ed542a742" exitCode=0 Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.499594 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerDied","Data":"37dc28f551f9db6ffb052a7a69754f617535fa9f02cee7492942ac3ed542a742"} Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.499628 4696 scope.go:117] "RemoveContainer" containerID="06727ab615d47d51e86a72ca06fb67dd4f0d596dbb88f8ea56eca5339ae5cc33" Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.504941 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-56b7788dc4-f928g" event={"ID":"3bd18ca4-e084-4fe1-85e3-7319ddd703ff","Type":"ContainerStarted","Data":"d1a7ffafb58d39990fe86f1227e98335a6ab7302575f7cdb6a37a59090097110"} Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.505235 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.505310 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.523470 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7c894bbd86-9phxx" podStartSLOduration=7.522825928 podStartE2EDuration="12.523449604s" podCreationTimestamp="2026-03-21 08:47:48 +0000 UTC" firstStartedPulling="2026-03-21 08:47:54.336593546 +0000 UTC m=+1208.457474259" lastFinishedPulling="2026-03-21 08:47:59.337217222 +0000 UTC m=+1213.458097935" observedRunningTime="2026-03-21 08:48:00.512367244 +0000 UTC m=+1214.633247957" watchObservedRunningTime="2026-03-21 08:48:00.523449604 +0000 UTC m=+1214.644330317" Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.550106 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-756466fd6c-6vdmf" podStartSLOduration=7.55084714 podStartE2EDuration="12.550082437s" podCreationTimestamp="2026-03-21 08:47:48 +0000 UTC" firstStartedPulling="2026-03-21 08:47:54.337872652 +0000 UTC m=+1208.458753375" lastFinishedPulling="2026-03-21 08:47:59.337107959 +0000 UTC m=+1213.457988672" observedRunningTime="2026-03-21 08:48:00.530867191 +0000 UTC m=+1214.651747904" watchObservedRunningTime="2026-03-21 08:48:00.550082437 +0000 UTC m=+1214.670963150" Mar 21 08:48:00 crc kubenswrapper[4696]: I0321 08:48:00.565198 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-56b7788dc4-f928g" podStartSLOduration=5.565181859 podStartE2EDuration="5.565181859s" podCreationTimestamp="2026-03-21 08:47:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:48:00.552455724 +0000 UTC m=+1214.673336437" watchObservedRunningTime="2026-03-21 08:48:00.565181859 +0000 UTC m=+1214.686062572" Mar 21 08:48:01 crc kubenswrapper[4696]: I0321 08:48:01.074836 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568048-j276n"] Mar 21 08:48:02 crc kubenswrapper[4696]: I0321 08:48:02.836289 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:48:03 crc kubenswrapper[4696]: I0321 08:48:03.445880 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:48:03 crc kubenswrapper[4696]: I0321 08:48:03.529102 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7dd88fbdb6-nnxhg" Mar 21 08:48:03 crc kubenswrapper[4696]: I0321 08:48:03.541940 4696 generic.go:334] "Generic (PLEG): container finished" podID="a70d1c8a-2db3-4ade-af61-fc8225262c18" containerID="499e153025457892a2d3dd1ae7d23b19b1557d52982f153cdf92e437caa1cd13" exitCode=0 Mar 21 08:48:03 crc kubenswrapper[4696]: I0321 08:48:03.542061 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-n5tdf" event={"ID":"a70d1c8a-2db3-4ade-af61-fc8225262c18","Type":"ContainerDied","Data":"499e153025457892a2d3dd1ae7d23b19b1557d52982f153cdf92e437caa1cd13"} Mar 21 08:48:03 crc kubenswrapper[4696]: I0321 08:48:03.547785 4696 generic.go:334] "Generic (PLEG): container finished" podID="d73897de-7ac5-4991-aabf-b36cc6ef424d" containerID="d6a2c9083ae133bb5944b16dc225cccbf9e3a5f9ea56291a7f5dd48fc657a5ad" exitCode=0 Mar 21 08:48:03 crc kubenswrapper[4696]: I0321 08:48:03.548714 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-8r5xw" event={"ID":"d73897de-7ac5-4991-aabf-b36cc6ef424d","Type":"ContainerDied","Data":"d6a2c9083ae133bb5944b16dc225cccbf9e3a5f9ea56291a7f5dd48fc657a5ad"} Mar 21 08:48:03 crc kubenswrapper[4696]: I0321 08:48:03.595759 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6b5744ff58-x5lj7"] Mar 21 08:48:03 crc kubenswrapper[4696]: I0321 08:48:03.595976 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6b5744ff58-x5lj7" podUID="876c7f89-b54b-4bcc-8220-33df4b204390" containerName="barbican-api-log" containerID="cri-o://5ea1280d4db7c529c5b86b33c1e58f42230d8eb4ccbf68af77facb55ec027f7f" gracePeriod=30 Mar 21 08:48:03 crc kubenswrapper[4696]: I0321 08:48:03.596251 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6b5744ff58-x5lj7" podUID="876c7f89-b54b-4bcc-8220-33df4b204390" containerName="barbican-api" containerID="cri-o://be0507642215e219f86a3c60e4e54ca5611563c572f345e8fd5b997b4ed7b90d" gracePeriod=30 Mar 21 08:48:03 crc kubenswrapper[4696]: I0321 08:48:03.607161 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6b5744ff58-x5lj7" podUID="876c7f89-b54b-4bcc-8220-33df4b204390" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.184:9311/healthcheck\": EOF" Mar 21 08:48:04 crc kubenswrapper[4696]: I0321 08:48:04.219036 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:48:04 crc kubenswrapper[4696]: I0321 08:48:04.292527 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-6mxl5"] Mar 21 08:48:04 crc kubenswrapper[4696]: I0321 08:48:04.292791 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" podUID="66c5da52-ed29-4222-8867-ac05aac6a382" containerName="dnsmasq-dns" containerID="cri-o://e26cba575b106923c7da050f84d362a3bb8475f3b104a4d2d5e7361c6fa84e77" gracePeriod=10 Mar 21 08:48:04 crc kubenswrapper[4696]: I0321 08:48:04.604523 4696 generic.go:334] "Generic (PLEG): container finished" podID="876c7f89-b54b-4bcc-8220-33df4b204390" containerID="5ea1280d4db7c529c5b86b33c1e58f42230d8eb4ccbf68af77facb55ec027f7f" exitCode=143 Mar 21 08:48:04 crc kubenswrapper[4696]: I0321 08:48:04.604590 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b5744ff58-x5lj7" event={"ID":"876c7f89-b54b-4bcc-8220-33df4b204390","Type":"ContainerDied","Data":"5ea1280d4db7c529c5b86b33c1e58f42230d8eb4ccbf68af77facb55ec027f7f"} Mar 21 08:48:04 crc kubenswrapper[4696]: E0321 08:48:04.929449 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66c5da52_ed29_4222_8867_ac05aac6a382.slice/crio-conmon-e26cba575b106923c7da050f84d362a3bb8475f3b104a4d2d5e7361c6fa84e77.scope\": RecentStats: unable to find data in memory cache]" Mar 21 08:48:05 crc kubenswrapper[4696]: I0321 08:48:05.616643 4696 generic.go:334] "Generic (PLEG): container finished" podID="66c5da52-ed29-4222-8867-ac05aac6a382" containerID="e26cba575b106923c7da050f84d362a3bb8475f3b104a4d2d5e7361c6fa84e77" exitCode=0 Mar 21 08:48:05 crc kubenswrapper[4696]: I0321 08:48:05.616718 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" event={"ID":"66c5da52-ed29-4222-8867-ac05aac6a382","Type":"ContainerDied","Data":"e26cba575b106923c7da050f84d362a3bb8475f3b104a4d2d5e7361c6fa84e77"} Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.020515 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.119015 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a70d1c8a-2db3-4ade-af61-fc8225262c18-etc-machine-id\") pod \"a70d1c8a-2db3-4ade-af61-fc8225262c18\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.119085 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a70d1c8a-2db3-4ade-af61-fc8225262c18-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a70d1c8a-2db3-4ade-af61-fc8225262c18" (UID: "a70d1c8a-2db3-4ade-af61-fc8225262c18"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.119357 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-scripts\") pod \"a70d1c8a-2db3-4ade-af61-fc8225262c18\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.119503 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpdc5\" (UniqueName: \"kubernetes.io/projected/a70d1c8a-2db3-4ade-af61-fc8225262c18-kube-api-access-dpdc5\") pod \"a70d1c8a-2db3-4ade-af61-fc8225262c18\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.119594 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-combined-ca-bundle\") pod \"a70d1c8a-2db3-4ade-af61-fc8225262c18\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.119667 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-db-sync-config-data\") pod \"a70d1c8a-2db3-4ade-af61-fc8225262c18\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.119704 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-config-data\") pod \"a70d1c8a-2db3-4ade-af61-fc8225262c18\" (UID: \"a70d1c8a-2db3-4ade-af61-fc8225262c18\") " Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.120306 4696 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a70d1c8a-2db3-4ade-af61-fc8225262c18-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.130270 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a70d1c8a-2db3-4ade-af61-fc8225262c18" (UID: "a70d1c8a-2db3-4ade-af61-fc8225262c18"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.132978 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a70d1c8a-2db3-4ade-af61-fc8225262c18-kube-api-access-dpdc5" (OuterVolumeSpecName: "kube-api-access-dpdc5") pod "a70d1c8a-2db3-4ade-af61-fc8225262c18" (UID: "a70d1c8a-2db3-4ade-af61-fc8225262c18"). InnerVolumeSpecName "kube-api-access-dpdc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.133110 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-scripts" (OuterVolumeSpecName: "scripts") pod "a70d1c8a-2db3-4ade-af61-fc8225262c18" (UID: "a70d1c8a-2db3-4ade-af61-fc8225262c18"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.142684 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" podUID="66c5da52-ed29-4222-8867-ac05aac6a382" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.175:5353: connect: connection refused" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.183901 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a70d1c8a-2db3-4ade-af61-fc8225262c18" (UID: "a70d1c8a-2db3-4ade-af61-fc8225262c18"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.222076 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.222108 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpdc5\" (UniqueName: \"kubernetes.io/projected/a70d1c8a-2db3-4ade-af61-fc8225262c18-kube-api-access-dpdc5\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.222117 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.222127 4696 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.230684 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-config-data" (OuterVolumeSpecName: "config-data") pod "a70d1c8a-2db3-4ade-af61-fc8225262c18" (UID: "a70d1c8a-2db3-4ade-af61-fc8225262c18"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.256906 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5d77f68d84-8hrbc" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.325746 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a70d1c8a-2db3-4ade-af61-fc8225262c18-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.555300 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-65c4b447c5-fstxp"] Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.555544 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-65c4b447c5-fstxp" podUID="19b4b8b0-047b-48bf-9c9b-70fe162cf5bd" containerName="neutron-api" containerID="cri-o://34604a8d20dd45769c938fc4e612566bd5eb98466fd47a1560d6276fec96a001" gracePeriod=30 Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.556172 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-65c4b447c5-fstxp" podUID="19b4b8b0-047b-48bf-9c9b-70fe162cf5bd" containerName="neutron-httpd" containerID="cri-o://5f83ec04b39227800d19bc750797e0e8b8ffaaee0a89e3f3298d96b0a49de970" gracePeriod=30 Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.605452 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-69b5b45b6f-7lhn7"] Mar 21 08:48:07 crc kubenswrapper[4696]: E0321 08:48:07.605877 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a70d1c8a-2db3-4ade-af61-fc8225262c18" containerName="cinder-db-sync" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.605892 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="a70d1c8a-2db3-4ade-af61-fc8225262c18" containerName="cinder-db-sync" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.606092 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="a70d1c8a-2db3-4ade-af61-fc8225262c18" containerName="cinder-db-sync" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.607379 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.650562 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69b5b45b6f-7lhn7"] Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.662603 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-65c4b447c5-fstxp" podUID="19b4b8b0-047b-48bf-9c9b-70fe162cf5bd" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.177:9696/\": EOF" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.668173 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-n5tdf" event={"ID":"a70d1c8a-2db3-4ade-af61-fc8225262c18","Type":"ContainerDied","Data":"30a2231543f7973200c0970b434c4ad623190a2282d35287b73a1da3e28caf22"} Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.668200 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-n5tdf" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.668210 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30a2231543f7973200c0970b434c4ad623190a2282d35287b73a1da3e28caf22" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.682506 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerStarted","Data":"25b5c4ade4d0fb1258c68b7e200638ed962f1ba1ce964274a9e7589cd0a163c2"} Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.735498 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1157f1b6-6891-43eb-93fc-e8e66691359a-combined-ca-bundle\") pod \"neutron-69b5b45b6f-7lhn7\" (UID: \"1157f1b6-6891-43eb-93fc-e8e66691359a\") " pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.735566 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1157f1b6-6891-43eb-93fc-e8e66691359a-public-tls-certs\") pod \"neutron-69b5b45b6f-7lhn7\" (UID: \"1157f1b6-6891-43eb-93fc-e8e66691359a\") " pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.735636 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzz57\" (UniqueName: \"kubernetes.io/projected/1157f1b6-6891-43eb-93fc-e8e66691359a-kube-api-access-gzz57\") pod \"neutron-69b5b45b6f-7lhn7\" (UID: \"1157f1b6-6891-43eb-93fc-e8e66691359a\") " pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.735660 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1157f1b6-6891-43eb-93fc-e8e66691359a-ovndb-tls-certs\") pod \"neutron-69b5b45b6f-7lhn7\" (UID: \"1157f1b6-6891-43eb-93fc-e8e66691359a\") " pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.735719 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1157f1b6-6891-43eb-93fc-e8e66691359a-config\") pod \"neutron-69b5b45b6f-7lhn7\" (UID: \"1157f1b6-6891-43eb-93fc-e8e66691359a\") " pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.735756 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1157f1b6-6891-43eb-93fc-e8e66691359a-internal-tls-certs\") pod \"neutron-69b5b45b6f-7lhn7\" (UID: \"1157f1b6-6891-43eb-93fc-e8e66691359a\") " pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.735784 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1157f1b6-6891-43eb-93fc-e8e66691359a-httpd-config\") pod \"neutron-69b5b45b6f-7lhn7\" (UID: \"1157f1b6-6891-43eb-93fc-e8e66691359a\") " pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.837514 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1157f1b6-6891-43eb-93fc-e8e66691359a-combined-ca-bundle\") pod \"neutron-69b5b45b6f-7lhn7\" (UID: \"1157f1b6-6891-43eb-93fc-e8e66691359a\") " pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.837566 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1157f1b6-6891-43eb-93fc-e8e66691359a-public-tls-certs\") pod \"neutron-69b5b45b6f-7lhn7\" (UID: \"1157f1b6-6891-43eb-93fc-e8e66691359a\") " pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.837623 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzz57\" (UniqueName: \"kubernetes.io/projected/1157f1b6-6891-43eb-93fc-e8e66691359a-kube-api-access-gzz57\") pod \"neutron-69b5b45b6f-7lhn7\" (UID: \"1157f1b6-6891-43eb-93fc-e8e66691359a\") " pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.837643 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1157f1b6-6891-43eb-93fc-e8e66691359a-ovndb-tls-certs\") pod \"neutron-69b5b45b6f-7lhn7\" (UID: \"1157f1b6-6891-43eb-93fc-e8e66691359a\") " pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.837688 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1157f1b6-6891-43eb-93fc-e8e66691359a-config\") pod \"neutron-69b5b45b6f-7lhn7\" (UID: \"1157f1b6-6891-43eb-93fc-e8e66691359a\") " pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.837716 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1157f1b6-6891-43eb-93fc-e8e66691359a-internal-tls-certs\") pod \"neutron-69b5b45b6f-7lhn7\" (UID: \"1157f1b6-6891-43eb-93fc-e8e66691359a\") " pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.837737 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1157f1b6-6891-43eb-93fc-e8e66691359a-httpd-config\") pod \"neutron-69b5b45b6f-7lhn7\" (UID: \"1157f1b6-6891-43eb-93fc-e8e66691359a\") " pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.842705 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1157f1b6-6891-43eb-93fc-e8e66691359a-public-tls-certs\") pod \"neutron-69b5b45b6f-7lhn7\" (UID: \"1157f1b6-6891-43eb-93fc-e8e66691359a\") " pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.843120 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1157f1b6-6891-43eb-93fc-e8e66691359a-httpd-config\") pod \"neutron-69b5b45b6f-7lhn7\" (UID: \"1157f1b6-6891-43eb-93fc-e8e66691359a\") " pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.844197 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1157f1b6-6891-43eb-93fc-e8e66691359a-config\") pod \"neutron-69b5b45b6f-7lhn7\" (UID: \"1157f1b6-6891-43eb-93fc-e8e66691359a\") " pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.845489 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1157f1b6-6891-43eb-93fc-e8e66691359a-internal-tls-certs\") pod \"neutron-69b5b45b6f-7lhn7\" (UID: \"1157f1b6-6891-43eb-93fc-e8e66691359a\") " pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.850258 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1157f1b6-6891-43eb-93fc-e8e66691359a-combined-ca-bundle\") pod \"neutron-69b5b45b6f-7lhn7\" (UID: \"1157f1b6-6891-43eb-93fc-e8e66691359a\") " pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.852356 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1157f1b6-6891-43eb-93fc-e8e66691359a-ovndb-tls-certs\") pod \"neutron-69b5b45b6f-7lhn7\" (UID: \"1157f1b6-6891-43eb-93fc-e8e66691359a\") " pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.854289 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzz57\" (UniqueName: \"kubernetes.io/projected/1157f1b6-6891-43eb-93fc-e8e66691359a-kube-api-access-gzz57\") pod \"neutron-69b5b45b6f-7lhn7\" (UID: \"1157f1b6-6891-43eb-93fc-e8e66691359a\") " pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:07 crc kubenswrapper[4696]: I0321 08:48:07.935873 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.029008 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6b5744ff58-x5lj7" podUID="876c7f89-b54b-4bcc-8220-33df4b204390" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.184:9311/healthcheck\": read tcp 10.217.0.2:48020->10.217.0.184:9311: read: connection reset by peer" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.040970 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6b5744ff58-x5lj7" podUID="876c7f89-b54b-4bcc-8220-33df4b204390" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.184:9311/healthcheck\": read tcp 10.217.0.2:48008->10.217.0.184:9311: read: connection reset by peer" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.041623 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6b5744ff58-x5lj7" podUID="876c7f89-b54b-4bcc-8220-33df4b204390" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.184:9311/healthcheck\": dial tcp 10.217.0.184:9311: connect: connection refused" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.353682 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-8r5xw" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.420275 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 21 08:48:08 crc kubenswrapper[4696]: E0321 08:48:08.422517 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d73897de-7ac5-4991-aabf-b36cc6ef424d" containerName="cloudkitty-storageinit" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.422649 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d73897de-7ac5-4991-aabf-b36cc6ef424d" containerName="cloudkitty-storageinit" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.423072 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="d73897de-7ac5-4991-aabf-b36cc6ef424d" containerName="cloudkitty-storageinit" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.454107 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.461546 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.463801 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d73897de-7ac5-4991-aabf-b36cc6ef424d-combined-ca-bundle\") pod \"d73897de-7ac5-4991-aabf-b36cc6ef424d\" (UID: \"d73897de-7ac5-4991-aabf-b36cc6ef424d\") " Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.487425 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d73897de-7ac5-4991-aabf-b36cc6ef424d-scripts\") pod \"d73897de-7ac5-4991-aabf-b36cc6ef424d\" (UID: \"d73897de-7ac5-4991-aabf-b36cc6ef424d\") " Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.487478 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d73897de-7ac5-4991-aabf-b36cc6ef424d-config-data\") pod \"d73897de-7ac5-4991-aabf-b36cc6ef424d\" (UID: \"d73897de-7ac5-4991-aabf-b36cc6ef424d\") " Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.487577 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d73897de-7ac5-4991-aabf-b36cc6ef424d-certs\") pod \"d73897de-7ac5-4991-aabf-b36cc6ef424d\" (UID: \"d73897de-7ac5-4991-aabf-b36cc6ef424d\") " Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.487606 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lchv\" (UniqueName: \"kubernetes.io/projected/d73897de-7ac5-4991-aabf-b36cc6ef424d-kube-api-access-5lchv\") pod \"d73897de-7ac5-4991-aabf-b36cc6ef424d\" (UID: \"d73897de-7ac5-4991-aabf-b36cc6ef424d\") " Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.503979 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d73897de-7ac5-4991-aabf-b36cc6ef424d-scripts" (OuterVolumeSpecName: "scripts") pod "d73897de-7ac5-4991-aabf-b36cc6ef424d" (UID: "d73897de-7ac5-4991-aabf-b36cc6ef424d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.504443 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-92r5d" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.504730 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.505807 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.506091 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.525730 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d73897de-7ac5-4991-aabf-b36cc6ef424d-kube-api-access-5lchv" (OuterVolumeSpecName: "kube-api-access-5lchv") pod "d73897de-7ac5-4991-aabf-b36cc6ef424d" (UID: "d73897de-7ac5-4991-aabf-b36cc6ef424d"). InnerVolumeSpecName "kube-api-access-5lchv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.532105 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d73897de-7ac5-4991-aabf-b36cc6ef424d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d73897de-7ac5-4991-aabf-b36cc6ef424d" (UID: "d73897de-7ac5-4991-aabf-b36cc6ef424d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.552863 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d73897de-7ac5-4991-aabf-b36cc6ef424d-certs" (OuterVolumeSpecName: "certs") pod "d73897de-7ac5-4991-aabf-b36cc6ef424d" (UID: "d73897de-7ac5-4991-aabf-b36cc6ef424d"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.585198 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-hrt4g"] Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.587003 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.588120 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.590584 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/33133b7f-8c96-4813-b1ab-2a40acb2cd84-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.590617 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.590674 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb7kr\" (UniqueName: \"kubernetes.io/projected/33133b7f-8c96-4813-b1ab-2a40acb2cd84-kube-api-access-hb7kr\") pod \"cinder-scheduler-0\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.590803 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-scripts\") pod \"cinder-scheduler-0\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.590903 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-config-data\") pod \"cinder-scheduler-0\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.590923 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.590988 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d73897de-7ac5-4991-aabf-b36cc6ef424d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.590999 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d73897de-7ac5-4991-aabf-b36cc6ef424d-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.591008 4696 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d73897de-7ac5-4991-aabf-b36cc6ef424d-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.591015 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lchv\" (UniqueName: \"kubernetes.io/projected/d73897de-7ac5-4991-aabf-b36cc6ef424d-kube-api-access-5lchv\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.591907 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-hrt4g"] Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.592848 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.598650 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.604910 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.682027 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d73897de-7ac5-4991-aabf-b36cc6ef424d-config-data" (OuterVolumeSpecName: "config-data") pod "d73897de-7ac5-4991-aabf-b36cc6ef424d" (UID: "d73897de-7ac5-4991-aabf-b36cc6ef424d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.694872 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck42j\" (UniqueName: \"kubernetes.io/projected/08575052-484c-47dd-91a4-bf5e182b8221-kube-api-access-ck42j\") pod \"dnsmasq-dns-5c9776ccc5-hrt4g\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.694934 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-config-data-custom\") pod \"cinder-api-0\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.694956 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-hrt4g\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.695006 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb7kr\" (UniqueName: \"kubernetes.io/projected/33133b7f-8c96-4813-b1ab-2a40acb2cd84-kube-api-access-hb7kr\") pod \"cinder-scheduler-0\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.695026 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-config-data\") pod \"cinder-api-0\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.695040 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e341e7e5-528c-42cb-b3df-725843538f37-logs\") pod \"cinder-api-0\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.695205 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-hrt4g\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.695227 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.695305 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-config\") pod \"dnsmasq-dns-5c9776ccc5-hrt4g\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.695332 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-scripts\") pod \"cinder-scheduler-0\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.695355 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-config-data\") pod \"cinder-scheduler-0\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.695374 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.695421 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-scripts\") pod \"cinder-api-0\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.695454 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cp2h\" (UniqueName: \"kubernetes.io/projected/e341e7e5-528c-42cb-b3df-725843538f37-kube-api-access-4cp2h\") pod \"cinder-api-0\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.695470 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e341e7e5-528c-42cb-b3df-725843538f37-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.695496 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/33133b7f-8c96-4813-b1ab-2a40acb2cd84-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.695516 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-hrt4g\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.695539 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.695559 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-hrt4g\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.695626 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d73897de-7ac5-4991-aabf-b36cc6ef424d-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.698585 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/33133b7f-8c96-4813-b1ab-2a40acb2cd84-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.708350 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.719642 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.720458 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-8r5xw" event={"ID":"d73897de-7ac5-4991-aabf-b36cc6ef424d","Type":"ContainerDied","Data":"8f426eb0e0c6a64f2e9272acebaaa0ce580d65fda3623c1726c9f20f26a79fdb"} Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.720528 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f426eb0e0c6a64f2e9272acebaaa0ce580d65fda3623c1726c9f20f26a79fdb" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.720638 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-8r5xw" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.736858 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-config-data\") pod \"cinder-scheduler-0\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.739870 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-scripts\") pod \"cinder-scheduler-0\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.743023 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb7kr\" (UniqueName: \"kubernetes.io/projected/33133b7f-8c96-4813-b1ab-2a40acb2cd84-kube-api-access-hb7kr\") pod \"cinder-scheduler-0\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.760778 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568048-j276n" event={"ID":"41361f63-e76c-42d2-ae18-32c5e772773a","Type":"ContainerStarted","Data":"44fc37f73359e7ade3cdbf3269ad5e42945bfe971c440e7e62868bc4a542c90c"} Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.775257 4696 generic.go:334] "Generic (PLEG): container finished" podID="876c7f89-b54b-4bcc-8220-33df4b204390" containerID="be0507642215e219f86a3c60e4e54ca5611563c572f345e8fd5b997b4ed7b90d" exitCode=0 Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.775352 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b5744ff58-x5lj7" event={"ID":"876c7f89-b54b-4bcc-8220-33df4b204390","Type":"ContainerDied","Data":"be0507642215e219f86a3c60e4e54ca5611563c572f345e8fd5b997b4ed7b90d"} Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.796486 4696 generic.go:334] "Generic (PLEG): container finished" podID="19b4b8b0-047b-48bf-9c9b-70fe162cf5bd" containerID="5f83ec04b39227800d19bc750797e0e8b8ffaaee0a89e3f3298d96b0a49de970" exitCode=0 Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.796552 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65c4b447c5-fstxp" event={"ID":"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd","Type":"ContainerDied","Data":"5f83ec04b39227800d19bc750797e0e8b8ffaaee0a89e3f3298d96b0a49de970"} Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.803042 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-config-data\") pod \"cinder-api-0\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.803086 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e341e7e5-528c-42cb-b3df-725843538f37-logs\") pod \"cinder-api-0\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.803244 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-hrt4g\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.803273 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.803350 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-config\") pod \"dnsmasq-dns-5c9776ccc5-hrt4g\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.803468 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-scripts\") pod \"cinder-api-0\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.803500 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cp2h\" (UniqueName: \"kubernetes.io/projected/e341e7e5-528c-42cb-b3df-725843538f37-kube-api-access-4cp2h\") pod \"cinder-api-0\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.803523 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e341e7e5-528c-42cb-b3df-725843538f37-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.803571 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-hrt4g\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.803602 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-hrt4g\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.803659 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck42j\" (UniqueName: \"kubernetes.io/projected/08575052-484c-47dd-91a4-bf5e182b8221-kube-api-access-ck42j\") pod \"dnsmasq-dns-5c9776ccc5-hrt4g\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.803693 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-config-data-custom\") pod \"cinder-api-0\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.803712 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-hrt4g\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.804679 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-hrt4g\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.805553 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e341e7e5-528c-42cb-b3df-725843538f37-logs\") pod \"cinder-api-0\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.807445 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e341e7e5-528c-42cb-b3df-725843538f37-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.808057 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-hrt4g\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.808728 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-config\") pod \"dnsmasq-dns-5c9776ccc5-hrt4g\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.809976 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-hrt4g\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.812270 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-config-data\") pod \"cinder-api-0\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.812463 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-hrt4g\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.821009 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.824670 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-config-data-custom\") pod \"cinder-api-0\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.825025 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-scripts\") pod \"cinder-api-0\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.827351 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cp2h\" (UniqueName: \"kubernetes.io/projected/e341e7e5-528c-42cb-b3df-725843538f37-kube-api-access-4cp2h\") pod \"cinder-api-0\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " pod="openstack/cinder-api-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.829482 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck42j\" (UniqueName: \"kubernetes.io/projected/08575052-484c-47dd-91a4-bf5e182b8221-kube-api-access-ck42j\") pod \"dnsmasq-dns-5c9776ccc5-hrt4g\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.908673 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.926251 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:08 crc kubenswrapper[4696]: I0321 08:48:08.966657 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.142556 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.225470 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-dns-svc\") pod \"66c5da52-ed29-4222-8867-ac05aac6a382\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.225763 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-ovsdbserver-sb\") pod \"66c5da52-ed29-4222-8867-ac05aac6a382\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.226228 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z98vc\" (UniqueName: \"kubernetes.io/projected/66c5da52-ed29-4222-8867-ac05aac6a382-kube-api-access-z98vc\") pod \"66c5da52-ed29-4222-8867-ac05aac6a382\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.226270 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-ovsdbserver-nb\") pod \"66c5da52-ed29-4222-8867-ac05aac6a382\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.226331 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-config\") pod \"66c5da52-ed29-4222-8867-ac05aac6a382\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.226374 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-dns-swift-storage-0\") pod \"66c5da52-ed29-4222-8867-ac05aac6a382\" (UID: \"66c5da52-ed29-4222-8867-ac05aac6a382\") " Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.236771 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66c5da52-ed29-4222-8867-ac05aac6a382-kube-api-access-z98vc" (OuterVolumeSpecName: "kube-api-access-z98vc") pod "66c5da52-ed29-4222-8867-ac05aac6a382" (UID: "66c5da52-ed29-4222-8867-ac05aac6a382"). InnerVolumeSpecName "kube-api-access-z98vc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.332527 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z98vc\" (UniqueName: \"kubernetes.io/projected/66c5da52-ed29-4222-8867-ac05aac6a382-kube-api-access-z98vc\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.364669 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "66c5da52-ed29-4222-8867-ac05aac6a382" (UID: "66c5da52-ed29-4222-8867-ac05aac6a382"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.375763 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.383010 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "66c5da52-ed29-4222-8867-ac05aac6a382" (UID: "66c5da52-ed29-4222-8867-ac05aac6a382"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.393002 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "66c5da52-ed29-4222-8867-ac05aac6a382" (UID: "66c5da52-ed29-4222-8867-ac05aac6a382"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.411465 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "66c5da52-ed29-4222-8867-ac05aac6a382" (UID: "66c5da52-ed29-4222-8867-ac05aac6a382"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.434639 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/876c7f89-b54b-4bcc-8220-33df4b204390-combined-ca-bundle\") pod \"876c7f89-b54b-4bcc-8220-33df4b204390\" (UID: \"876c7f89-b54b-4bcc-8220-33df4b204390\") " Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.434847 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/876c7f89-b54b-4bcc-8220-33df4b204390-config-data-custom\") pod \"876c7f89-b54b-4bcc-8220-33df4b204390\" (UID: \"876c7f89-b54b-4bcc-8220-33df4b204390\") " Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.434904 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/876c7f89-b54b-4bcc-8220-33df4b204390-config-data\") pod \"876c7f89-b54b-4bcc-8220-33df4b204390\" (UID: \"876c7f89-b54b-4bcc-8220-33df4b204390\") " Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.434959 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/876c7f89-b54b-4bcc-8220-33df4b204390-logs\") pod \"876c7f89-b54b-4bcc-8220-33df4b204390\" (UID: \"876c7f89-b54b-4bcc-8220-33df4b204390\") " Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.435058 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxzvr\" (UniqueName: \"kubernetes.io/projected/876c7f89-b54b-4bcc-8220-33df4b204390-kube-api-access-dxzvr\") pod \"876c7f89-b54b-4bcc-8220-33df4b204390\" (UID: \"876c7f89-b54b-4bcc-8220-33df4b204390\") " Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.435428 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.435439 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.435450 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.435459 4696 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.445912 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/876c7f89-b54b-4bcc-8220-33df4b204390-logs" (OuterVolumeSpecName: "logs") pod "876c7f89-b54b-4bcc-8220-33df4b204390" (UID: "876c7f89-b54b-4bcc-8220-33df4b204390"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.449539 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-config" (OuterVolumeSpecName: "config") pod "66c5da52-ed29-4222-8867-ac05aac6a382" (UID: "66c5da52-ed29-4222-8867-ac05aac6a382"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.466077 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/876c7f89-b54b-4bcc-8220-33df4b204390-kube-api-access-dxzvr" (OuterVolumeSpecName: "kube-api-access-dxzvr") pod "876c7f89-b54b-4bcc-8220-33df4b204390" (UID: "876c7f89-b54b-4bcc-8220-33df4b204390"). InnerVolumeSpecName "kube-api-access-dxzvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.467548 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/876c7f89-b54b-4bcc-8220-33df4b204390-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "876c7f89-b54b-4bcc-8220-33df4b204390" (UID: "876c7f89-b54b-4bcc-8220-33df4b204390"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.537571 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/876c7f89-b54b-4bcc-8220-33df4b204390-logs\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.537598 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxzvr\" (UniqueName: \"kubernetes.io/projected/876c7f89-b54b-4bcc-8220-33df4b204390-kube-api-access-dxzvr\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.537613 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66c5da52-ed29-4222-8867-ac05aac6a382-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.537740 4696 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/876c7f89-b54b-4bcc-8220-33df4b204390-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.561178 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/876c7f89-b54b-4bcc-8220-33df4b204390-config-data" (OuterVolumeSpecName: "config-data") pod "876c7f89-b54b-4bcc-8220-33df4b204390" (UID: "876c7f89-b54b-4bcc-8220-33df4b204390"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.610893 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69b5b45b6f-7lhn7"] Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.624083 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/876c7f89-b54b-4bcc-8220-33df4b204390-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "876c7f89-b54b-4bcc-8220-33df4b204390" (UID: "876c7f89-b54b-4bcc-8220-33df4b204390"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.645180 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/876c7f89-b54b-4bcc-8220-33df4b204390-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.645211 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/876c7f89-b54b-4bcc-8220-33df4b204390-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.731907 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-proc-0"] Mar 21 08:48:09 crc kubenswrapper[4696]: E0321 08:48:09.732519 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66c5da52-ed29-4222-8867-ac05aac6a382" containerName="dnsmasq-dns" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.732530 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="66c5da52-ed29-4222-8867-ac05aac6a382" containerName="dnsmasq-dns" Mar 21 08:48:09 crc kubenswrapper[4696]: E0321 08:48:09.732542 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="876c7f89-b54b-4bcc-8220-33df4b204390" containerName="barbican-api-log" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.732548 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="876c7f89-b54b-4bcc-8220-33df4b204390" containerName="barbican-api-log" Mar 21 08:48:09 crc kubenswrapper[4696]: E0321 08:48:09.732566 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="876c7f89-b54b-4bcc-8220-33df4b204390" containerName="barbican-api" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.732572 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="876c7f89-b54b-4bcc-8220-33df4b204390" containerName="barbican-api" Mar 21 08:48:09 crc kubenswrapper[4696]: E0321 08:48:09.732588 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66c5da52-ed29-4222-8867-ac05aac6a382" containerName="init" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.732593 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="66c5da52-ed29-4222-8867-ac05aac6a382" containerName="init" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.732771 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="876c7f89-b54b-4bcc-8220-33df4b204390" containerName="barbican-api" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.732804 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="876c7f89-b54b-4bcc-8220-33df4b204390" containerName="barbican-api-log" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.732811 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="66c5da52-ed29-4222-8867-ac05aac6a382" containerName="dnsmasq-dns" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.733478 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.745164 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-ff6v6" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.745396 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.745547 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.763055 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.763275 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-proc-config-data" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.794249 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.798953 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-65c4b447c5-fstxp" podUID="19b4b8b0-047b-48bf-9c9b-70fe162cf5bd" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.177:9696/\": dial tcp 10.217.0.177:9696: connect: connection refused" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.876314 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-scripts\") pod \"cloudkitty-proc-0\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.876383 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-config-data\") pod \"cloudkitty-proc-0\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.876460 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwrmx\" (UniqueName: \"kubernetes.io/projected/ba009f97-23a1-4296-8f34-aeb08fcc6b80-kube-api-access-pwrmx\") pod \"cloudkitty-proc-0\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.876510 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/ba009f97-23a1-4296-8f34-aeb08fcc6b80-certs\") pod \"cloudkitty-proc-0\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.876540 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.876659 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.937167 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-hrt4g"] Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.972130 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69b5b45b6f-7lhn7" event={"ID":"1157f1b6-6891-43eb-93fc-e8e66691359a","Type":"ContainerStarted","Data":"f752fe2cacd37f98f2ca87c1ffc2592155fe73aa056b193edb6b6e089e0d8e6f"} Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.987071 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67bdc55879-kc4xr"] Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.988513 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67bdc55879-kc4xr"] Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.988586 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.989574 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-scripts\") pod \"cloudkitty-proc-0\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.989620 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-config-data\") pod \"cloudkitty-proc-0\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.989667 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwrmx\" (UniqueName: \"kubernetes.io/projected/ba009f97-23a1-4296-8f34-aeb08fcc6b80-kube-api-access-pwrmx\") pod \"cloudkitty-proc-0\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.989701 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/ba009f97-23a1-4296-8f34-aeb08fcc6b80-certs\") pod \"cloudkitty-proc-0\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.989725 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:09 crc kubenswrapper[4696]: I0321 08:48:09.989798 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.001433 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.028315 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-scripts\") pod \"cloudkitty-proc-0\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.059110 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.064649 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.066235 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/ba009f97-23a1-4296-8f34-aeb08fcc6b80-certs\") pod \"cloudkitty-proc-0\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.075428 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwrmx\" (UniqueName: \"kubernetes.io/projected/ba009f97-23a1-4296-8f34-aeb08fcc6b80-kube-api-access-pwrmx\") pod \"cloudkitty-proc-0\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.091354 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b5744ff58-x5lj7" event={"ID":"876c7f89-b54b-4bcc-8220-33df4b204390","Type":"ContainerDied","Data":"ee9d5ed650ca674ed29b53bc011acaf13cc16e9e5f48e0a71f58c8ad2a67cadc"} Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.091558 4696 scope.go:117] "RemoveContainer" containerID="be0507642215e219f86a3c60e4e54ca5611563c572f345e8fd5b997b4ed7b90d" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.091958 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b5744ff58-x5lj7" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.117697 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-config-data\") pod \"cloudkitty-proc-0\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.136504 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-ovsdbserver-nb\") pod \"dnsmasq-dns-67bdc55879-kc4xr\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.136644 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-dns-swift-storage-0\") pod \"dnsmasq-dns-67bdc55879-kc4xr\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.136752 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-ovsdbserver-sb\") pod \"dnsmasq-dns-67bdc55879-kc4xr\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.136833 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gw2p\" (UniqueName: \"kubernetes.io/projected/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-kube-api-access-7gw2p\") pod \"dnsmasq-dns-67bdc55879-kc4xr\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.136853 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-dns-svc\") pod \"dnsmasq-dns-67bdc55879-kc4xr\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.136956 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-config\") pod \"dnsmasq-dns-67bdc55879-kc4xr\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.160768 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.172363 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbd625f3-5fcc-48f5-9022-df08fdc01887","Type":"ContainerStarted","Data":"4ef06e4a275bb3a92e419224699d016d81119d7d043d36daa87dba6035758ec0"} Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.172586 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbd625f3-5fcc-48f5-9022-df08fdc01887" containerName="ceilometer-central-agent" containerID="cri-o://7d3f92e8692b9fc0fa147cacc12efb0194a77da1acc10ad757b8fb7c002a9a62" gracePeriod=30 Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.174147 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.174203 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbd625f3-5fcc-48f5-9022-df08fdc01887" containerName="proxy-httpd" containerID="cri-o://4ef06e4a275bb3a92e419224699d016d81119d7d043d36daa87dba6035758ec0" gracePeriod=30 Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.174221 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbd625f3-5fcc-48f5-9022-df08fdc01887" containerName="ceilometer-notification-agent" containerID="cri-o://0c5965d091d4ecbb20a816b1558e00f85fbc6379914b0c2e493d1c70c053bb4d" gracePeriod=30 Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.174292 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbd625f3-5fcc-48f5-9022-df08fdc01887" containerName="sg-core" containerID="cri-o://2e4eed46e1a4de0167e86432becd21c8cb3965e91eaa6b63e2084ca18ed3b13e" gracePeriod=30 Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.175059 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-api-0"] Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.179772 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.183240 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-api-config-data" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.233234 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" event={"ID":"66c5da52-ed29-4222-8867-ac05aac6a382","Type":"ContainerDied","Data":"c5a2b89c8b11dd92388e049e0363d4ef438dacf9e3bad408b15e6ae38124c69e"} Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.234172 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-6mxl5" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.246725 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-dns-swift-storage-0\") pod \"dnsmasq-dns-67bdc55879-kc4xr\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.246793 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-scripts\") pod \"cloudkitty-api-0\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.246849 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-config-data\") pod \"cloudkitty-api-0\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.246909 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/a17438b6-e92b-4e6e-90dd-6843714adc35-certs\") pod \"cloudkitty-api-0\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.246971 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-ovsdbserver-sb\") pod \"dnsmasq-dns-67bdc55879-kc4xr\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.247032 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gw2p\" (UniqueName: \"kubernetes.io/projected/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-kube-api-access-7gw2p\") pod \"dnsmasq-dns-67bdc55879-kc4xr\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.247058 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-dns-svc\") pod \"dnsmasq-dns-67bdc55879-kc4xr\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.247089 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.247112 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a17438b6-e92b-4e6e-90dd-6843714adc35-logs\") pod \"cloudkitty-api-0\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.247147 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.247218 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkvfm\" (UniqueName: \"kubernetes.io/projected/a17438b6-e92b-4e6e-90dd-6843714adc35-kube-api-access-nkvfm\") pod \"cloudkitty-api-0\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.247290 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-config\") pod \"dnsmasq-dns-67bdc55879-kc4xr\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.247334 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-ovsdbserver-nb\") pod \"dnsmasq-dns-67bdc55879-kc4xr\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.248538 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-ovsdbserver-nb\") pod \"dnsmasq-dns-67bdc55879-kc4xr\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.249243 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-dns-swift-storage-0\") pod \"dnsmasq-dns-67bdc55879-kc4xr\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.249743 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.250394 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-ovsdbserver-sb\") pod \"dnsmasq-dns-67bdc55879-kc4xr\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.250659 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-config\") pod \"dnsmasq-dns-67bdc55879-kc4xr\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.260071 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-dns-svc\") pod \"dnsmasq-dns-67bdc55879-kc4xr\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.274649 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-hrt4g"] Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.275237 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.285808305 podStartE2EDuration="1m8.275218522s" podCreationTimestamp="2026-03-21 08:47:02 +0000 UTC" firstStartedPulling="2026-03-21 08:47:04.546470349 +0000 UTC m=+1158.667351062" lastFinishedPulling="2026-03-21 08:48:08.535880566 +0000 UTC m=+1222.656761279" observedRunningTime="2026-03-21 08:48:10.229861346 +0000 UTC m=+1224.350742069" watchObservedRunningTime="2026-03-21 08:48:10.275218522 +0000 UTC m=+1224.396099235" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.298151 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gw2p\" (UniqueName: \"kubernetes.io/projected/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-kube-api-access-7gw2p\") pod \"dnsmasq-dns-67bdc55879-kc4xr\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.349357 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-scripts\") pod \"cloudkitty-api-0\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.349710 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-config-data\") pod \"cloudkitty-api-0\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.349775 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/a17438b6-e92b-4e6e-90dd-6843714adc35-certs\") pod \"cloudkitty-api-0\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.349924 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.349953 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a17438b6-e92b-4e6e-90dd-6843714adc35-logs\") pod \"cloudkitty-api-0\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.350004 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.350077 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkvfm\" (UniqueName: \"kubernetes.io/projected/a17438b6-e92b-4e6e-90dd-6843714adc35-kube-api-access-nkvfm\") pod \"cloudkitty-api-0\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.351759 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a17438b6-e92b-4e6e-90dd-6843714adc35-logs\") pod \"cloudkitty-api-0\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.355360 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.356046 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-config-data\") pod \"cloudkitty-api-0\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.356607 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.356774 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-scripts\") pod \"cloudkitty-api-0\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.364474 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.373310 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/a17438b6-e92b-4e6e-90dd-6843714adc35-certs\") pod \"cloudkitty-api-0\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.373651 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkvfm\" (UniqueName: \"kubernetes.io/projected/a17438b6-e92b-4e6e-90dd-6843714adc35-kube-api-access-nkvfm\") pod \"cloudkitty-api-0\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.429887 4696 scope.go:117] "RemoveContainer" containerID="5ea1280d4db7c529c5b86b33c1e58f42230d8eb4ccbf68af77facb55ec027f7f" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.495304 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.497003 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.497383 4696 scope.go:117] "RemoveContainer" containerID="e26cba575b106923c7da050f84d362a3bb8475f3b104a4d2d5e7361c6fa84e77" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.578044 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-6mxl5"] Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.578086 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-6mxl5"] Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.580990 4696 scope.go:117] "RemoveContainer" containerID="cfffd805cedc9becdabd33fdd877d3f0ffbf1f5ac7d9f680703abc9d2b48c5a7" Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.581376 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6b5744ff58-x5lj7"] Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.592842 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6b5744ff58-x5lj7"] Mar 21 08:48:10 crc kubenswrapper[4696]: I0321 08:48:10.727121 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.255744 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"33133b7f-8c96-4813-b1ab-2a40acb2cd84","Type":"ContainerStarted","Data":"5a2aa71528c016a1e6faab82e79b728928036f78624c31bdce4e952c9b4bf5f0"} Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.262133 4696 generic.go:334] "Generic (PLEG): container finished" podID="41361f63-e76c-42d2-ae18-32c5e772773a" containerID="b296773d278d198bdcff27110fa8b8aee8a0160dbe3d66d891888e4b8c7a2532" exitCode=0 Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.262214 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568048-j276n" event={"ID":"41361f63-e76c-42d2-ae18-32c5e772773a","Type":"ContainerDied","Data":"b296773d278d198bdcff27110fa8b8aee8a0160dbe3d66d891888e4b8c7a2532"} Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.265792 4696 generic.go:334] "Generic (PLEG): container finished" podID="08575052-484c-47dd-91a4-bf5e182b8221" containerID="060b07f07d611363965ee7d344167713a568b62df40a58db81742750b59672d0" exitCode=0 Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.265918 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" event={"ID":"08575052-484c-47dd-91a4-bf5e182b8221","Type":"ContainerDied","Data":"060b07f07d611363965ee7d344167713a568b62df40a58db81742750b59672d0"} Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.265952 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" event={"ID":"08575052-484c-47dd-91a4-bf5e182b8221","Type":"ContainerStarted","Data":"044b98d62014c2190d33360cc0398e9326239884c3551b847dfd8de2bb2f75e0"} Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.273831 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69b5b45b6f-7lhn7" event={"ID":"1157f1b6-6891-43eb-93fc-e8e66691359a","Type":"ContainerStarted","Data":"cbccab350da41419f86ab459186d369bae6bb6850842fe14ba39bfad36b751de"} Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.273883 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69b5b45b6f-7lhn7" event={"ID":"1157f1b6-6891-43eb-93fc-e8e66691359a","Type":"ContainerStarted","Data":"d76c4a8591a1ea7dc3bb68e8bb8a0f59e5038303e8cb6ec8015522dc8b305eea"} Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.274974 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.276597 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e341e7e5-528c-42cb-b3df-725843538f37","Type":"ContainerStarted","Data":"fac4eac9b00ce530becf1893befa4354e10c44fbb09845db832ce793e84d44c1"} Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.300726 4696 generic.go:334] "Generic (PLEG): container finished" podID="dbd625f3-5fcc-48f5-9022-df08fdc01887" containerID="4ef06e4a275bb3a92e419224699d016d81119d7d043d36daa87dba6035758ec0" exitCode=0 Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.301289 4696 generic.go:334] "Generic (PLEG): container finished" podID="dbd625f3-5fcc-48f5-9022-df08fdc01887" containerID="2e4eed46e1a4de0167e86432becd21c8cb3965e91eaa6b63e2084ca18ed3b13e" exitCode=2 Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.301304 4696 generic.go:334] "Generic (PLEG): container finished" podID="dbd625f3-5fcc-48f5-9022-df08fdc01887" containerID="7d3f92e8692b9fc0fa147cacc12efb0194a77da1acc10ad757b8fb7c002a9a62" exitCode=0 Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.301376 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbd625f3-5fcc-48f5-9022-df08fdc01887","Type":"ContainerDied","Data":"4ef06e4a275bb3a92e419224699d016d81119d7d043d36daa87dba6035758ec0"} Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.301409 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbd625f3-5fcc-48f5-9022-df08fdc01887","Type":"ContainerDied","Data":"2e4eed46e1a4de0167e86432becd21c8cb3965e91eaa6b63e2084ca18ed3b13e"} Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.301426 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbd625f3-5fcc-48f5-9022-df08fdc01887","Type":"ContainerDied","Data":"7d3f92e8692b9fc0fa147cacc12efb0194a77da1acc10ad757b8fb7c002a9a62"} Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.316556 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67bdc55879-kc4xr"] Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.336981 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"ba009f97-23a1-4296-8f34-aeb08fcc6b80","Type":"ContainerStarted","Data":"0ce2ebece7506499e7f3d5dc03ee27110d15a1383a50dd10f282d5cab8967d45"} Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.359259 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-69b5b45b6f-7lhn7" podStartSLOduration=4.359236722 podStartE2EDuration="4.359236722s" podCreationTimestamp="2026-03-21 08:48:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:48:11.327224599 +0000 UTC m=+1225.448105322" watchObservedRunningTime="2026-03-21 08:48:11.359236722 +0000 UTC m=+1225.480117455" Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.459275 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Mar 21 08:48:11 crc kubenswrapper[4696]: W0321 08:48:11.471794 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda17438b6_e92b_4e6e_90dd_6843714adc35.slice/crio-eb4aa5abef4754dcdae58575b39384b5119cd9cd8be3b7691349867e4a4d4718 WatchSource:0}: Error finding container eb4aa5abef4754dcdae58575b39384b5119cd9cd8be3b7691349867e4a4d4718: Status 404 returned error can't find the container with id eb4aa5abef4754dcdae58575b39384b5119cd9cd8be3b7691349867e4a4d4718 Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.653175 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.758026 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.807363 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-ovsdbserver-nb\") pod \"08575052-484c-47dd-91a4-bf5e182b8221\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.807441 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-dns-svc\") pod \"08575052-484c-47dd-91a4-bf5e182b8221\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.807505 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-ovsdbserver-sb\") pod \"08575052-484c-47dd-91a4-bf5e182b8221\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.807550 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-dns-swift-storage-0\") pod \"08575052-484c-47dd-91a4-bf5e182b8221\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.807605 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ck42j\" (UniqueName: \"kubernetes.io/projected/08575052-484c-47dd-91a4-bf5e182b8221-kube-api-access-ck42j\") pod \"08575052-484c-47dd-91a4-bf5e182b8221\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.807658 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-config\") pod \"08575052-484c-47dd-91a4-bf5e182b8221\" (UID: \"08575052-484c-47dd-91a4-bf5e182b8221\") " Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.837144 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08575052-484c-47dd-91a4-bf5e182b8221-kube-api-access-ck42j" (OuterVolumeSpecName: "kube-api-access-ck42j") pod "08575052-484c-47dd-91a4-bf5e182b8221" (UID: "08575052-484c-47dd-91a4-bf5e182b8221"). InnerVolumeSpecName "kube-api-access-ck42j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.884507 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "08575052-484c-47dd-91a4-bf5e182b8221" (UID: "08575052-484c-47dd-91a4-bf5e182b8221"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.889699 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "08575052-484c-47dd-91a4-bf5e182b8221" (UID: "08575052-484c-47dd-91a4-bf5e182b8221"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.910175 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-config" (OuterVolumeSpecName: "config") pod "08575052-484c-47dd-91a4-bf5e182b8221" (UID: "08575052-484c-47dd-91a4-bf5e182b8221"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.910505 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ck42j\" (UniqueName: \"kubernetes.io/projected/08575052-484c-47dd-91a4-bf5e182b8221-kube-api-access-ck42j\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.910535 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.910546 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.910554 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.917208 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "08575052-484c-47dd-91a4-bf5e182b8221" (UID: "08575052-484c-47dd-91a4-bf5e182b8221"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:48:11 crc kubenswrapper[4696]: I0321 08:48:11.918310 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "08575052-484c-47dd-91a4-bf5e182b8221" (UID: "08575052-484c-47dd-91a4-bf5e182b8221"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:48:12 crc kubenswrapper[4696]: I0321 08:48:12.012329 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:12 crc kubenswrapper[4696]: I0321 08:48:12.012359 4696 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/08575052-484c-47dd-91a4-bf5e182b8221-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:12 crc kubenswrapper[4696]: I0321 08:48:12.396294 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e341e7e5-528c-42cb-b3df-725843538f37","Type":"ContainerStarted","Data":"f564e73050da5326436e14b1e6ee5f2fbc582d0841f8e51443f69e3d5914feb3"} Mar 21 08:48:12 crc kubenswrapper[4696]: I0321 08:48:12.451713 4696 generic.go:334] "Generic (PLEG): container finished" podID="19b4b8b0-047b-48bf-9c9b-70fe162cf5bd" containerID="34604a8d20dd45769c938fc4e612566bd5eb98466fd47a1560d6276fec96a001" exitCode=0 Mar 21 08:48:12 crc kubenswrapper[4696]: I0321 08:48:12.452166 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65c4b447c5-fstxp" event={"ID":"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd","Type":"ContainerDied","Data":"34604a8d20dd45769c938fc4e612566bd5eb98466fd47a1560d6276fec96a001"} Mar 21 08:48:12 crc kubenswrapper[4696]: I0321 08:48:12.457849 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" event={"ID":"08575052-484c-47dd-91a4-bf5e182b8221","Type":"ContainerDied","Data":"044b98d62014c2190d33360cc0398e9326239884c3551b847dfd8de2bb2f75e0"} Mar 21 08:48:12 crc kubenswrapper[4696]: I0321 08:48:12.457901 4696 scope.go:117] "RemoveContainer" containerID="060b07f07d611363965ee7d344167713a568b62df40a58db81742750b59672d0" Mar 21 08:48:12 crc kubenswrapper[4696]: I0321 08:48:12.457974 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-hrt4g" Mar 21 08:48:12 crc kubenswrapper[4696]: I0321 08:48:12.471587 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"a17438b6-e92b-4e6e-90dd-6843714adc35","Type":"ContainerStarted","Data":"18367b044502afe3de0a6c99ba84b504c44a31dc6a00ff4630ce9973c48c5660"} Mar 21 08:48:12 crc kubenswrapper[4696]: I0321 08:48:12.471638 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"a17438b6-e92b-4e6e-90dd-6843714adc35","Type":"ContainerStarted","Data":"eb4aa5abef4754dcdae58575b39384b5119cd9cd8be3b7691349867e4a4d4718"} Mar 21 08:48:12 crc kubenswrapper[4696]: I0321 08:48:12.474266 4696 generic.go:334] "Generic (PLEG): container finished" podID="2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1" containerID="02b2c669c853f02bcbddcb5cce028bbad7b960d0821dd2cd4d12972bd0670824" exitCode=0 Mar 21 08:48:12 crc kubenswrapper[4696]: I0321 08:48:12.474484 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" event={"ID":"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1","Type":"ContainerDied","Data":"02b2c669c853f02bcbddcb5cce028bbad7b960d0821dd2cd4d12972bd0670824"} Mar 21 08:48:12 crc kubenswrapper[4696]: I0321 08:48:12.474522 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" event={"ID":"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1","Type":"ContainerStarted","Data":"257c3c64522b08c742fb41cd2edcc8e6bd95b7b76a449c671fce1f6ec1963b0b"} Mar 21 08:48:12 crc kubenswrapper[4696]: I0321 08:48:12.580707 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66c5da52-ed29-4222-8867-ac05aac6a382" path="/var/lib/kubelet/pods/66c5da52-ed29-4222-8867-ac05aac6a382/volumes" Mar 21 08:48:12 crc kubenswrapper[4696]: I0321 08:48:12.581359 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="876c7f89-b54b-4bcc-8220-33df4b204390" path="/var/lib/kubelet/pods/876c7f89-b54b-4bcc-8220-33df4b204390/volumes" Mar 21 08:48:12 crc kubenswrapper[4696]: I0321 08:48:12.581968 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-hrt4g"] Mar 21 08:48:12 crc kubenswrapper[4696]: I0321 08:48:12.597503 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-hrt4g"] Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.499069 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568048-j276n" event={"ID":"41361f63-e76c-42d2-ae18-32c5e772773a","Type":"ContainerDied","Data":"44fc37f73359e7ade3cdbf3269ad5e42945bfe971c440e7e62868bc4a542c90c"} Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.499529 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44fc37f73359e7ade3cdbf3269ad5e42945bfe971c440e7e62868bc4a542c90c" Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.503650 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"a17438b6-e92b-4e6e-90dd-6843714adc35","Type":"ContainerStarted","Data":"3f8919fe290a02f8091f7acbae0079823adb80b458b523abc25f4c7d3b864940"} Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.505160 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-api-0" Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.509375 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e341e7e5-528c-42cb-b3df-725843538f37","Type":"ContainerStarted","Data":"1dad4f74a4cf2b0e3c0ab3d7d81e837b523309be69d052ce2ca128608539edb1"} Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.509509 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e341e7e5-528c-42cb-b3df-725843538f37" containerName="cinder-api-log" containerID="cri-o://f564e73050da5326436e14b1e6ee5f2fbc582d0841f8e51443f69e3d5914feb3" gracePeriod=30 Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.509545 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.509870 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e341e7e5-528c-42cb-b3df-725843538f37" containerName="cinder-api" containerID="cri-o://1dad4f74a4cf2b0e3c0ab3d7d81e837b523309be69d052ce2ca128608539edb1" gracePeriod=30 Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.511147 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"33133b7f-8c96-4813-b1ab-2a40acb2cd84","Type":"ContainerStarted","Data":"cc664ad69a991d8883bbae1c6ec054b7b3b665f8bd2f4a92612ba533f977da0f"} Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.513597 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65c4b447c5-fstxp" event={"ID":"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd","Type":"ContainerDied","Data":"de19182eab2d7b11b3b6fc702051f49c65be082ba99d7df0443f6f1bf8947a4a"} Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.513641 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de19182eab2d7b11b3b6fc702051f49c65be082ba99d7df0443f6f1bf8947a4a" Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.588285 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-api-0" podStartSLOduration=3.588264544 podStartE2EDuration="3.588264544s" podCreationTimestamp="2026-03-21 08:48:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:48:13.569519072 +0000 UTC m=+1227.690399785" watchObservedRunningTime="2026-03-21 08:48:13.588264544 +0000 UTC m=+1227.709145257" Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.656865 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.656845538 podStartE2EDuration="5.656845538s" podCreationTimestamp="2026-03-21 08:48:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:48:13.641176041 +0000 UTC m=+1227.762056754" watchObservedRunningTime="2026-03-21 08:48:13.656845538 +0000 UTC m=+1227.777726251" Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.820007 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.843129 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568048-j276n" Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.874507 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-combined-ca-bundle\") pod \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.874599 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-httpd-config\") pod \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.874644 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-public-tls-certs\") pod \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.874696 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gj22\" (UniqueName: \"kubernetes.io/projected/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-kube-api-access-5gj22\") pod \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.874738 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6vxm\" (UniqueName: \"kubernetes.io/projected/41361f63-e76c-42d2-ae18-32c5e772773a-kube-api-access-r6vxm\") pod \"41361f63-e76c-42d2-ae18-32c5e772773a\" (UID: \"41361f63-e76c-42d2-ae18-32c5e772773a\") " Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.874857 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-config\") pod \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.874916 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-ovndb-tls-certs\") pod \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.874941 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-internal-tls-certs\") pod \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\" (UID: \"19b4b8b0-047b-48bf-9c9b-70fe162cf5bd\") " Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.880281 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "19b4b8b0-047b-48bf-9c9b-70fe162cf5bd" (UID: "19b4b8b0-047b-48bf-9c9b-70fe162cf5bd"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.883660 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-kube-api-access-5gj22" (OuterVolumeSpecName: "kube-api-access-5gj22") pod "19b4b8b0-047b-48bf-9c9b-70fe162cf5bd" (UID: "19b4b8b0-047b-48bf-9c9b-70fe162cf5bd"). InnerVolumeSpecName "kube-api-access-5gj22". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.894061 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41361f63-e76c-42d2-ae18-32c5e772773a-kube-api-access-r6vxm" (OuterVolumeSpecName: "kube-api-access-r6vxm") pod "41361f63-e76c-42d2-ae18-32c5e772773a" (UID: "41361f63-e76c-42d2-ae18-32c5e772773a"). InnerVolumeSpecName "kube-api-access-r6vxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.977476 4696 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.977509 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gj22\" (UniqueName: \"kubernetes.io/projected/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-kube-api-access-5gj22\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:13 crc kubenswrapper[4696]: I0321 08:48:13.977520 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6vxm\" (UniqueName: \"kubernetes.io/projected/41361f63-e76c-42d2-ae18-32c5e772773a-kube-api-access-r6vxm\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.087249 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "19b4b8b0-047b-48bf-9c9b-70fe162cf5bd" (UID: "19b4b8b0-047b-48bf-9c9b-70fe162cf5bd"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.090429 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-config" (OuterVolumeSpecName: "config") pod "19b4b8b0-047b-48bf-9c9b-70fe162cf5bd" (UID: "19b4b8b0-047b-48bf-9c9b-70fe162cf5bd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.108057 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "19b4b8b0-047b-48bf-9c9b-70fe162cf5bd" (UID: "19b4b8b0-047b-48bf-9c9b-70fe162cf5bd"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.109299 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19b4b8b0-047b-48bf-9c9b-70fe162cf5bd" (UID: "19b4b8b0-047b-48bf-9c9b-70fe162cf5bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.123126 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "19b4b8b0-047b-48bf-9c9b-70fe162cf5bd" (UID: "19b4b8b0-047b-48bf-9c9b-70fe162cf5bd"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.178713 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.181284 4696 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.181311 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.181322 4696 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.181332 4696 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.181341 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.282894 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-combined-ca-bundle\") pod \"e341e7e5-528c-42cb-b3df-725843538f37\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.282983 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e341e7e5-528c-42cb-b3df-725843538f37-logs\") pod \"e341e7e5-528c-42cb-b3df-725843538f37\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.283053 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-scripts\") pod \"e341e7e5-528c-42cb-b3df-725843538f37\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.283116 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cp2h\" (UniqueName: \"kubernetes.io/projected/e341e7e5-528c-42cb-b3df-725843538f37-kube-api-access-4cp2h\") pod \"e341e7e5-528c-42cb-b3df-725843538f37\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.283179 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e341e7e5-528c-42cb-b3df-725843538f37-etc-machine-id\") pod \"e341e7e5-528c-42cb-b3df-725843538f37\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.283267 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-config-data\") pod \"e341e7e5-528c-42cb-b3df-725843538f37\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.283323 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-config-data-custom\") pod \"e341e7e5-528c-42cb-b3df-725843538f37\" (UID: \"e341e7e5-528c-42cb-b3df-725843538f37\") " Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.283321 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e341e7e5-528c-42cb-b3df-725843538f37-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e341e7e5-528c-42cb-b3df-725843538f37" (UID: "e341e7e5-528c-42cb-b3df-725843538f37"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.283539 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e341e7e5-528c-42cb-b3df-725843538f37-logs" (OuterVolumeSpecName: "logs") pod "e341e7e5-528c-42cb-b3df-725843538f37" (UID: "e341e7e5-528c-42cb-b3df-725843538f37"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.283987 4696 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e341e7e5-528c-42cb-b3df-725843538f37-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.284007 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e341e7e5-528c-42cb-b3df-725843538f37-logs\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.286757 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-scripts" (OuterVolumeSpecName: "scripts") pod "e341e7e5-528c-42cb-b3df-725843538f37" (UID: "e341e7e5-528c-42cb-b3df-725843538f37"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.289457 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e341e7e5-528c-42cb-b3df-725843538f37" (UID: "e341e7e5-528c-42cb-b3df-725843538f37"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.299146 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e341e7e5-528c-42cb-b3df-725843538f37-kube-api-access-4cp2h" (OuterVolumeSpecName: "kube-api-access-4cp2h") pod "e341e7e5-528c-42cb-b3df-725843538f37" (UID: "e341e7e5-528c-42cb-b3df-725843538f37"). InnerVolumeSpecName "kube-api-access-4cp2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.314200 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e341e7e5-528c-42cb-b3df-725843538f37" (UID: "e341e7e5-528c-42cb-b3df-725843538f37"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.330510 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.355515 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-config-data" (OuterVolumeSpecName: "config-data") pod "e341e7e5-528c-42cb-b3df-725843538f37" (UID: "e341e7e5-528c-42cb-b3df-725843538f37"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.386372 4696 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.386406 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.386420 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.386428 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cp2h\" (UniqueName: \"kubernetes.io/projected/e341e7e5-528c-42cb-b3df-725843538f37-kube-api-access-4cp2h\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.386438 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e341e7e5-528c-42cb-b3df-725843538f37-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.578517 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08575052-484c-47dd-91a4-bf5e182b8221" path="/var/lib/kubelet/pods/08575052-484c-47dd-91a4-bf5e182b8221/volumes" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.590509 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"ba009f97-23a1-4296-8f34-aeb08fcc6b80","Type":"ContainerStarted","Data":"e563f58ed0293b5674d023311cc2edf051d98f1f64074f381b1c742fad20ec28"} Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.590554 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" event={"ID":"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1","Type":"ContainerStarted","Data":"ade47d5d8b5fc01f442812f9ccf8b8af5c438979e8d544094bdc36466086ef07"} Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.590884 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.599566 4696 generic.go:334] "Generic (PLEG): container finished" podID="e341e7e5-528c-42cb-b3df-725843538f37" containerID="1dad4f74a4cf2b0e3c0ab3d7d81e837b523309be69d052ce2ca128608539edb1" exitCode=0 Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.599595 4696 generic.go:334] "Generic (PLEG): container finished" podID="e341e7e5-528c-42cb-b3df-725843538f37" containerID="f564e73050da5326436e14b1e6ee5f2fbc582d0841f8e51443f69e3d5914feb3" exitCode=143 Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.604169 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e341e7e5-528c-42cb-b3df-725843538f37","Type":"ContainerDied","Data":"1dad4f74a4cf2b0e3c0ab3d7d81e837b523309be69d052ce2ca128608539edb1"} Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.604290 4696 scope.go:117] "RemoveContainer" containerID="1dad4f74a4cf2b0e3c0ab3d7d81e837b523309be69d052ce2ca128608539edb1" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.604242 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e341e7e5-528c-42cb-b3df-725843538f37","Type":"ContainerDied","Data":"f564e73050da5326436e14b1e6ee5f2fbc582d0841f8e51443f69e3d5914feb3"} Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.604699 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e341e7e5-528c-42cb-b3df-725843538f37","Type":"ContainerDied","Data":"fac4eac9b00ce530becf1893befa4354e10c44fbb09845db832ce793e84d44c1"} Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.604885 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.640221 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568048-j276n" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.647584 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"33133b7f-8c96-4813-b1ab-2a40acb2cd84","Type":"ContainerStarted","Data":"db14586774b84c9cfaa3cc665fa911d8ac8c18592c7f5dac9340bfddddc13b20"} Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.647694 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65c4b447c5-fstxp" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.664955 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-proc-0" podStartSLOduration=2.878982695 podStartE2EDuration="5.664933729s" podCreationTimestamp="2026-03-21 08:48:09 +0000 UTC" firstStartedPulling="2026-03-21 08:48:10.742024268 +0000 UTC m=+1224.862904981" lastFinishedPulling="2026-03-21 08:48:13.527975302 +0000 UTC m=+1227.648856015" observedRunningTime="2026-03-21 08:48:14.624129991 +0000 UTC m=+1228.745010704" watchObservedRunningTime="2026-03-21 08:48:14.664933729 +0000 UTC m=+1228.785814432" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.685580 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.685947 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" podStartSLOduration=5.685929755 podStartE2EDuration="5.685929755s" podCreationTimestamp="2026-03-21 08:48:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:48:14.664464557 +0000 UTC m=+1228.785345270" watchObservedRunningTime="2026-03-21 08:48:14.685929755 +0000 UTC m=+1228.806810458" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.742112 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.227234749 podStartE2EDuration="6.742087802s" podCreationTimestamp="2026-03-21 08:48:08 +0000 UTC" firstStartedPulling="2026-03-21 08:48:10.150682017 +0000 UTC m=+1224.271562730" lastFinishedPulling="2026-03-21 08:48:11.66553507 +0000 UTC m=+1225.786415783" observedRunningTime="2026-03-21 08:48:14.730936951 +0000 UTC m=+1228.851817654" watchObservedRunningTime="2026-03-21 08:48:14.742087802 +0000 UTC m=+1228.862968515" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.862272 4696 scope.go:117] "RemoveContainer" containerID="f564e73050da5326436e14b1e6ee5f2fbc582d0841f8e51443f69e3d5914feb3" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.887854 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.916447 4696 scope.go:117] "RemoveContainer" containerID="1dad4f74a4cf2b0e3c0ab3d7d81e837b523309be69d052ce2ca128608539edb1" Mar 21 08:48:14 crc kubenswrapper[4696]: E0321 08:48:14.916803 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dad4f74a4cf2b0e3c0ab3d7d81e837b523309be69d052ce2ca128608539edb1\": container with ID starting with 1dad4f74a4cf2b0e3c0ab3d7d81e837b523309be69d052ce2ca128608539edb1 not found: ID does not exist" containerID="1dad4f74a4cf2b0e3c0ab3d7d81e837b523309be69d052ce2ca128608539edb1" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.916883 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dad4f74a4cf2b0e3c0ab3d7d81e837b523309be69d052ce2ca128608539edb1"} err="failed to get container status \"1dad4f74a4cf2b0e3c0ab3d7d81e837b523309be69d052ce2ca128608539edb1\": rpc error: code = NotFound desc = could not find container \"1dad4f74a4cf2b0e3c0ab3d7d81e837b523309be69d052ce2ca128608539edb1\": container with ID starting with 1dad4f74a4cf2b0e3c0ab3d7d81e837b523309be69d052ce2ca128608539edb1 not found: ID does not exist" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.916908 4696 scope.go:117] "RemoveContainer" containerID="f564e73050da5326436e14b1e6ee5f2fbc582d0841f8e51443f69e3d5914feb3" Mar 21 08:48:14 crc kubenswrapper[4696]: E0321 08:48:14.917265 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f564e73050da5326436e14b1e6ee5f2fbc582d0841f8e51443f69e3d5914feb3\": container with ID starting with f564e73050da5326436e14b1e6ee5f2fbc582d0841f8e51443f69e3d5914feb3 not found: ID does not exist" containerID="f564e73050da5326436e14b1e6ee5f2fbc582d0841f8e51443f69e3d5914feb3" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.917287 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f564e73050da5326436e14b1e6ee5f2fbc582d0841f8e51443f69e3d5914feb3"} err="failed to get container status \"f564e73050da5326436e14b1e6ee5f2fbc582d0841f8e51443f69e3d5914feb3\": rpc error: code = NotFound desc = could not find container \"f564e73050da5326436e14b1e6ee5f2fbc582d0841f8e51443f69e3d5914feb3\": container with ID starting with f564e73050da5326436e14b1e6ee5f2fbc582d0841f8e51443f69e3d5914feb3 not found: ID does not exist" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.917300 4696 scope.go:117] "RemoveContainer" containerID="1dad4f74a4cf2b0e3c0ab3d7d81e837b523309be69d052ce2ca128608539edb1" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.917517 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dad4f74a4cf2b0e3c0ab3d7d81e837b523309be69d052ce2ca128608539edb1"} err="failed to get container status \"1dad4f74a4cf2b0e3c0ab3d7d81e837b523309be69d052ce2ca128608539edb1\": rpc error: code = NotFound desc = could not find container \"1dad4f74a4cf2b0e3c0ab3d7d81e837b523309be69d052ce2ca128608539edb1\": container with ID starting with 1dad4f74a4cf2b0e3c0ab3d7d81e837b523309be69d052ce2ca128608539edb1 not found: ID does not exist" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.917533 4696 scope.go:117] "RemoveContainer" containerID="f564e73050da5326436e14b1e6ee5f2fbc582d0841f8e51443f69e3d5914feb3" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.917954 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f564e73050da5326436e14b1e6ee5f2fbc582d0841f8e51443f69e3d5914feb3"} err="failed to get container status \"f564e73050da5326436e14b1e6ee5f2fbc582d0841f8e51443f69e3d5914feb3\": rpc error: code = NotFound desc = could not find container \"f564e73050da5326436e14b1e6ee5f2fbc582d0841f8e51443f69e3d5914feb3\": container with ID starting with f564e73050da5326436e14b1e6ee5f2fbc582d0841f8e51443f69e3d5914feb3 not found: ID does not exist" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.923884 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.944036 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-65c4b447c5-fstxp"] Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.988328 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-65c4b447c5-fstxp"] Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.998273 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 21 08:48:14 crc kubenswrapper[4696]: E0321 08:48:14.998694 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19b4b8b0-047b-48bf-9c9b-70fe162cf5bd" containerName="neutron-httpd" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.998711 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="19b4b8b0-047b-48bf-9c9b-70fe162cf5bd" containerName="neutron-httpd" Mar 21 08:48:14 crc kubenswrapper[4696]: E0321 08:48:14.998724 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08575052-484c-47dd-91a4-bf5e182b8221" containerName="init" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.998731 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="08575052-484c-47dd-91a4-bf5e182b8221" containerName="init" Mar 21 08:48:14 crc kubenswrapper[4696]: E0321 08:48:14.998755 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e341e7e5-528c-42cb-b3df-725843538f37" containerName="cinder-api-log" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.998761 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e341e7e5-528c-42cb-b3df-725843538f37" containerName="cinder-api-log" Mar 21 08:48:14 crc kubenswrapper[4696]: E0321 08:48:14.998786 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e341e7e5-528c-42cb-b3df-725843538f37" containerName="cinder-api" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.998792 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e341e7e5-528c-42cb-b3df-725843538f37" containerName="cinder-api" Mar 21 08:48:14 crc kubenswrapper[4696]: E0321 08:48:14.998802 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41361f63-e76c-42d2-ae18-32c5e772773a" containerName="oc" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.998807 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="41361f63-e76c-42d2-ae18-32c5e772773a" containerName="oc" Mar 21 08:48:14 crc kubenswrapper[4696]: E0321 08:48:14.998828 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19b4b8b0-047b-48bf-9c9b-70fe162cf5bd" containerName="neutron-api" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.998833 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="19b4b8b0-047b-48bf-9c9b-70fe162cf5bd" containerName="neutron-api" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.999019 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="08575052-484c-47dd-91a4-bf5e182b8221" containerName="init" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.999046 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="19b4b8b0-047b-48bf-9c9b-70fe162cf5bd" containerName="neutron-httpd" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.999054 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="e341e7e5-528c-42cb-b3df-725843538f37" containerName="cinder-api" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.999066 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="19b4b8b0-047b-48bf-9c9b-70fe162cf5bd" containerName="neutron-api" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.999076 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="41361f63-e76c-42d2-ae18-32c5e772773a" containerName="oc" Mar 21 08:48:14 crc kubenswrapper[4696]: I0321 08:48:14.999082 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="e341e7e5-528c-42cb-b3df-725843538f37" containerName="cinder-api-log" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.001371 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.006895 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.007036 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.007130 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.011567 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.030772 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568042-xb457"] Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.038940 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568042-xb457"] Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.112374 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d228678-e5cf-4da4-9a96-7ed39576c142-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.112415 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d228678-e5cf-4da4-9a96-7ed39576c142-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.112444 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d228678-e5cf-4da4-9a96-7ed39576c142-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.112479 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d228678-e5cf-4da4-9a96-7ed39576c142-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.112565 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d228678-e5cf-4da4-9a96-7ed39576c142-logs\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.112585 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2qv4\" (UniqueName: \"kubernetes.io/projected/8d228678-e5cf-4da4-9a96-7ed39576c142-kube-api-access-m2qv4\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.112611 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d228678-e5cf-4da4-9a96-7ed39576c142-config-data-custom\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.112667 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d228678-e5cf-4da4-9a96-7ed39576c142-scripts\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.112684 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d228678-e5cf-4da4-9a96-7ed39576c142-config-data\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.214499 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d228678-e5cf-4da4-9a96-7ed39576c142-scripts\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.215397 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d228678-e5cf-4da4-9a96-7ed39576c142-config-data\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.215454 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d228678-e5cf-4da4-9a96-7ed39576c142-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.215471 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d228678-e5cf-4da4-9a96-7ed39576c142-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.215500 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d228678-e5cf-4da4-9a96-7ed39576c142-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.216120 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d228678-e5cf-4da4-9a96-7ed39576c142-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.216209 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d228678-e5cf-4da4-9a96-7ed39576c142-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.216416 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d228678-e5cf-4da4-9a96-7ed39576c142-logs\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.216458 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2qv4\" (UniqueName: \"kubernetes.io/projected/8d228678-e5cf-4da4-9a96-7ed39576c142-kube-api-access-m2qv4\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.216542 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d228678-e5cf-4da4-9a96-7ed39576c142-config-data-custom\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.221557 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d228678-e5cf-4da4-9a96-7ed39576c142-logs\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.226058 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d228678-e5cf-4da4-9a96-7ed39576c142-config-data\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.229405 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d228678-e5cf-4da4-9a96-7ed39576c142-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.229876 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d228678-e5cf-4da4-9a96-7ed39576c142-config-data-custom\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.232303 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d228678-e5cf-4da4-9a96-7ed39576c142-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.240489 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d228678-e5cf-4da4-9a96-7ed39576c142-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.241331 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d228678-e5cf-4da4-9a96-7ed39576c142-scripts\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.246382 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2qv4\" (UniqueName: \"kubernetes.io/projected/8d228678-e5cf-4da4-9a96-7ed39576c142-kube-api-access-m2qv4\") pod \"cinder-api-0\" (UID: \"8d228678-e5cf-4da4-9a96-7ed39576c142\") " pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.331786 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.374975 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.522615 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbd625f3-5fcc-48f5-9022-df08fdc01887-log-httpd\") pod \"dbd625f3-5fcc-48f5-9022-df08fdc01887\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.522932 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-sg-core-conf-yaml\") pod \"dbd625f3-5fcc-48f5-9022-df08fdc01887\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.523094 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4crtt\" (UniqueName: \"kubernetes.io/projected/dbd625f3-5fcc-48f5-9022-df08fdc01887-kube-api-access-4crtt\") pod \"dbd625f3-5fcc-48f5-9022-df08fdc01887\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.523148 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-combined-ca-bundle\") pod \"dbd625f3-5fcc-48f5-9022-df08fdc01887\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.523172 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbd625f3-5fcc-48f5-9022-df08fdc01887-run-httpd\") pod \"dbd625f3-5fcc-48f5-9022-df08fdc01887\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.523242 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-scripts\") pod \"dbd625f3-5fcc-48f5-9022-df08fdc01887\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.523286 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-config-data\") pod \"dbd625f3-5fcc-48f5-9022-df08fdc01887\" (UID: \"dbd625f3-5fcc-48f5-9022-df08fdc01887\") " Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.527467 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbd625f3-5fcc-48f5-9022-df08fdc01887-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "dbd625f3-5fcc-48f5-9022-df08fdc01887" (UID: "dbd625f3-5fcc-48f5-9022-df08fdc01887"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.527793 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbd625f3-5fcc-48f5-9022-df08fdc01887-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "dbd625f3-5fcc-48f5-9022-df08fdc01887" (UID: "dbd625f3-5fcc-48f5-9022-df08fdc01887"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.530150 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbd625f3-5fcc-48f5-9022-df08fdc01887-kube-api-access-4crtt" (OuterVolumeSpecName: "kube-api-access-4crtt") pod "dbd625f3-5fcc-48f5-9022-df08fdc01887" (UID: "dbd625f3-5fcc-48f5-9022-df08fdc01887"). InnerVolumeSpecName "kube-api-access-4crtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.544904 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-scripts" (OuterVolumeSpecName: "scripts") pod "dbd625f3-5fcc-48f5-9022-df08fdc01887" (UID: "dbd625f3-5fcc-48f5-9022-df08fdc01887"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.552407 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "dbd625f3-5fcc-48f5-9022-df08fdc01887" (UID: "dbd625f3-5fcc-48f5-9022-df08fdc01887"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.626192 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4crtt\" (UniqueName: \"kubernetes.io/projected/dbd625f3-5fcc-48f5-9022-df08fdc01887-kube-api-access-4crtt\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.626222 4696 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbd625f3-5fcc-48f5-9022-df08fdc01887-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.626235 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.626247 4696 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbd625f3-5fcc-48f5-9022-df08fdc01887-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.626255 4696 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.635188 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dbd625f3-5fcc-48f5-9022-df08fdc01887" (UID: "dbd625f3-5fcc-48f5-9022-df08fdc01887"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.661135 4696 generic.go:334] "Generic (PLEG): container finished" podID="dbd625f3-5fcc-48f5-9022-df08fdc01887" containerID="0c5965d091d4ecbb20a816b1558e00f85fbc6379914b0c2e493d1c70c053bb4d" exitCode=0 Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.662086 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.662573 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbd625f3-5fcc-48f5-9022-df08fdc01887","Type":"ContainerDied","Data":"0c5965d091d4ecbb20a816b1558e00f85fbc6379914b0c2e493d1c70c053bb4d"} Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.662598 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbd625f3-5fcc-48f5-9022-df08fdc01887","Type":"ContainerDied","Data":"e0430d561aa504fb166fa0c66a1816a40fd58d7f009269cec7b22c7098f0ce92"} Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.662615 4696 scope.go:117] "RemoveContainer" containerID="4ef06e4a275bb3a92e419224699d016d81119d7d043d36daa87dba6035758ec0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.667750 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="a17438b6-e92b-4e6e-90dd-6843714adc35" containerName="cloudkitty-api-log" containerID="cri-o://18367b044502afe3de0a6c99ba84b504c44a31dc6a00ff4630ce9973c48c5660" gracePeriod=30 Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.667908 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="a17438b6-e92b-4e6e-90dd-6843714adc35" containerName="cloudkitty-api" containerID="cri-o://3f8919fe290a02f8091f7acbae0079823adb80b458b523abc25f4c7d3b864940" gracePeriod=30 Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.684027 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-config-data" (OuterVolumeSpecName: "config-data") pod "dbd625f3-5fcc-48f5-9022-df08fdc01887" (UID: "dbd625f3-5fcc-48f5-9022-df08fdc01887"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.728367 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.728413 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbd625f3-5fcc-48f5-9022-df08fdc01887-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.794181 4696 scope.go:117] "RemoveContainer" containerID="2e4eed46e1a4de0167e86432becd21c8cb3965e91eaa6b63e2084ca18ed3b13e" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.821937 4696 scope.go:117] "RemoveContainer" containerID="0c5965d091d4ecbb20a816b1558e00f85fbc6379914b0c2e493d1c70c053bb4d" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.851832 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.860190 4696 scope.go:117] "RemoveContainer" containerID="7d3f92e8692b9fc0fa147cacc12efb0194a77da1acc10ad757b8fb7c002a9a62" Mar 21 08:48:15 crc kubenswrapper[4696]: W0321 08:48:15.870592 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d228678_e5cf_4da4_9a96_7ed39576c142.slice/crio-cc0ff885d9078ca9dbd72b8d4086da743400cbb01ef9530975ee469345e849ad WatchSource:0}: Error finding container cc0ff885d9078ca9dbd72b8d4086da743400cbb01ef9530975ee469345e849ad: Status 404 returned error can't find the container with id cc0ff885d9078ca9dbd72b8d4086da743400cbb01ef9530975ee469345e849ad Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.893410 4696 scope.go:117] "RemoveContainer" containerID="4ef06e4a275bb3a92e419224699d016d81119d7d043d36daa87dba6035758ec0" Mar 21 08:48:15 crc kubenswrapper[4696]: E0321 08:48:15.893865 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ef06e4a275bb3a92e419224699d016d81119d7d043d36daa87dba6035758ec0\": container with ID starting with 4ef06e4a275bb3a92e419224699d016d81119d7d043d36daa87dba6035758ec0 not found: ID does not exist" containerID="4ef06e4a275bb3a92e419224699d016d81119d7d043d36daa87dba6035758ec0" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.893895 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ef06e4a275bb3a92e419224699d016d81119d7d043d36daa87dba6035758ec0"} err="failed to get container status \"4ef06e4a275bb3a92e419224699d016d81119d7d043d36daa87dba6035758ec0\": rpc error: code = NotFound desc = could not find container \"4ef06e4a275bb3a92e419224699d016d81119d7d043d36daa87dba6035758ec0\": container with ID starting with 4ef06e4a275bb3a92e419224699d016d81119d7d043d36daa87dba6035758ec0 not found: ID does not exist" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.893916 4696 scope.go:117] "RemoveContainer" containerID="2e4eed46e1a4de0167e86432becd21c8cb3965e91eaa6b63e2084ca18ed3b13e" Mar 21 08:48:15 crc kubenswrapper[4696]: E0321 08:48:15.894983 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e4eed46e1a4de0167e86432becd21c8cb3965e91eaa6b63e2084ca18ed3b13e\": container with ID starting with 2e4eed46e1a4de0167e86432becd21c8cb3965e91eaa6b63e2084ca18ed3b13e not found: ID does not exist" containerID="2e4eed46e1a4de0167e86432becd21c8cb3965e91eaa6b63e2084ca18ed3b13e" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.895036 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e4eed46e1a4de0167e86432becd21c8cb3965e91eaa6b63e2084ca18ed3b13e"} err="failed to get container status \"2e4eed46e1a4de0167e86432becd21c8cb3965e91eaa6b63e2084ca18ed3b13e\": rpc error: code = NotFound desc = could not find container \"2e4eed46e1a4de0167e86432becd21c8cb3965e91eaa6b63e2084ca18ed3b13e\": container with ID starting with 2e4eed46e1a4de0167e86432becd21c8cb3965e91eaa6b63e2084ca18ed3b13e not found: ID does not exist" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.895072 4696 scope.go:117] "RemoveContainer" containerID="0c5965d091d4ecbb20a816b1558e00f85fbc6379914b0c2e493d1c70c053bb4d" Mar 21 08:48:15 crc kubenswrapper[4696]: E0321 08:48:15.895416 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c5965d091d4ecbb20a816b1558e00f85fbc6379914b0c2e493d1c70c053bb4d\": container with ID starting with 0c5965d091d4ecbb20a816b1558e00f85fbc6379914b0c2e493d1c70c053bb4d not found: ID does not exist" containerID="0c5965d091d4ecbb20a816b1558e00f85fbc6379914b0c2e493d1c70c053bb4d" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.895443 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c5965d091d4ecbb20a816b1558e00f85fbc6379914b0c2e493d1c70c053bb4d"} err="failed to get container status \"0c5965d091d4ecbb20a816b1558e00f85fbc6379914b0c2e493d1c70c053bb4d\": rpc error: code = NotFound desc = could not find container \"0c5965d091d4ecbb20a816b1558e00f85fbc6379914b0c2e493d1c70c053bb4d\": container with ID starting with 0c5965d091d4ecbb20a816b1558e00f85fbc6379914b0c2e493d1c70c053bb4d not found: ID does not exist" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.895457 4696 scope.go:117] "RemoveContainer" containerID="7d3f92e8692b9fc0fa147cacc12efb0194a77da1acc10ad757b8fb7c002a9a62" Mar 21 08:48:15 crc kubenswrapper[4696]: E0321 08:48:15.895754 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d3f92e8692b9fc0fa147cacc12efb0194a77da1acc10ad757b8fb7c002a9a62\": container with ID starting with 7d3f92e8692b9fc0fa147cacc12efb0194a77da1acc10ad757b8fb7c002a9a62 not found: ID does not exist" containerID="7d3f92e8692b9fc0fa147cacc12efb0194a77da1acc10ad757b8fb7c002a9a62" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.895781 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d3f92e8692b9fc0fa147cacc12efb0194a77da1acc10ad757b8fb7c002a9a62"} err="failed to get container status \"7d3f92e8692b9fc0fa147cacc12efb0194a77da1acc10ad757b8fb7c002a9a62\": rpc error: code = NotFound desc = could not find container \"7d3f92e8692b9fc0fa147cacc12efb0194a77da1acc10ad757b8fb7c002a9a62\": container with ID starting with 7d3f92e8692b9fc0fa147cacc12efb0194a77da1acc10ad757b8fb7c002a9a62 not found: ID does not exist" Mar 21 08:48:15 crc kubenswrapper[4696]: I0321 08:48:15.993670 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.006198 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.028184 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:48:16 crc kubenswrapper[4696]: E0321 08:48:16.028846 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbd625f3-5fcc-48f5-9022-df08fdc01887" containerName="sg-core" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.028866 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbd625f3-5fcc-48f5-9022-df08fdc01887" containerName="sg-core" Mar 21 08:48:16 crc kubenswrapper[4696]: E0321 08:48:16.028902 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbd625f3-5fcc-48f5-9022-df08fdc01887" containerName="ceilometer-central-agent" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.028909 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbd625f3-5fcc-48f5-9022-df08fdc01887" containerName="ceilometer-central-agent" Mar 21 08:48:16 crc kubenswrapper[4696]: E0321 08:48:16.028918 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbd625f3-5fcc-48f5-9022-df08fdc01887" containerName="ceilometer-notification-agent" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.028924 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbd625f3-5fcc-48f5-9022-df08fdc01887" containerName="ceilometer-notification-agent" Mar 21 08:48:16 crc kubenswrapper[4696]: E0321 08:48:16.028935 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbd625f3-5fcc-48f5-9022-df08fdc01887" containerName="proxy-httpd" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.028940 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbd625f3-5fcc-48f5-9022-df08fdc01887" containerName="proxy-httpd" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.031141 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbd625f3-5fcc-48f5-9022-df08fdc01887" containerName="proxy-httpd" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.031175 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbd625f3-5fcc-48f5-9022-df08fdc01887" containerName="sg-core" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.031188 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbd625f3-5fcc-48f5-9022-df08fdc01887" containerName="ceilometer-notification-agent" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.031212 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbd625f3-5fcc-48f5-9022-df08fdc01887" containerName="ceilometer-central-agent" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.078891 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.079001 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.082465 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.083348 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.241709 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kznqc\" (UniqueName: \"kubernetes.io/projected/1e007714-252e-43e9-b281-323db9d580ad-kube-api-access-kznqc\") pod \"ceilometer-0\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.241783 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-config-data\") pod \"ceilometer-0\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.241848 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-scripts\") pod \"ceilometer-0\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.241871 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.241889 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.241919 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e007714-252e-43e9-b281-323db9d580ad-run-httpd\") pod \"ceilometer-0\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.241941 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e007714-252e-43e9-b281-323db9d580ad-log-httpd\") pod \"ceilometer-0\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.292574 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.343782 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kznqc\" (UniqueName: \"kubernetes.io/projected/1e007714-252e-43e9-b281-323db9d580ad-kube-api-access-kznqc\") pod \"ceilometer-0\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.343988 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-config-data\") pod \"ceilometer-0\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.344055 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-scripts\") pod \"ceilometer-0\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.344089 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.344115 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.344148 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e007714-252e-43e9-b281-323db9d580ad-run-httpd\") pod \"ceilometer-0\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.344183 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e007714-252e-43e9-b281-323db9d580ad-log-httpd\") pod \"ceilometer-0\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.344947 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e007714-252e-43e9-b281-323db9d580ad-log-httpd\") pod \"ceilometer-0\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.345325 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e007714-252e-43e9-b281-323db9d580ad-run-httpd\") pod \"ceilometer-0\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.349075 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.349760 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.350450 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-scripts\") pod \"ceilometer-0\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.351386 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-config-data\") pod \"ceilometer-0\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.364414 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kznqc\" (UniqueName: \"kubernetes.io/projected/1e007714-252e-43e9-b281-323db9d580ad-kube-api-access-kznqc\") pod \"ceilometer-0\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.402652 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.445055 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-config-data\") pod \"a17438b6-e92b-4e6e-90dd-6843714adc35\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.445118 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-config-data-custom\") pod \"a17438b6-e92b-4e6e-90dd-6843714adc35\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.445230 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a17438b6-e92b-4e6e-90dd-6843714adc35-logs\") pod \"a17438b6-e92b-4e6e-90dd-6843714adc35\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.445316 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/a17438b6-e92b-4e6e-90dd-6843714adc35-certs\") pod \"a17438b6-e92b-4e6e-90dd-6843714adc35\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.445344 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-scripts\") pod \"a17438b6-e92b-4e6e-90dd-6843714adc35\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.445462 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-combined-ca-bundle\") pod \"a17438b6-e92b-4e6e-90dd-6843714adc35\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.445527 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkvfm\" (UniqueName: \"kubernetes.io/projected/a17438b6-e92b-4e6e-90dd-6843714adc35-kube-api-access-nkvfm\") pod \"a17438b6-e92b-4e6e-90dd-6843714adc35\" (UID: \"a17438b6-e92b-4e6e-90dd-6843714adc35\") " Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.450462 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-scripts" (OuterVolumeSpecName: "scripts") pod "a17438b6-e92b-4e6e-90dd-6843714adc35" (UID: "a17438b6-e92b-4e6e-90dd-6843714adc35"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.451690 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a17438b6-e92b-4e6e-90dd-6843714adc35" (UID: "a17438b6-e92b-4e6e-90dd-6843714adc35"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.452035 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a17438b6-e92b-4e6e-90dd-6843714adc35-kube-api-access-nkvfm" (OuterVolumeSpecName: "kube-api-access-nkvfm") pod "a17438b6-e92b-4e6e-90dd-6843714adc35" (UID: "a17438b6-e92b-4e6e-90dd-6843714adc35"). InnerVolumeSpecName "kube-api-access-nkvfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.453656 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a17438b6-e92b-4e6e-90dd-6843714adc35-certs" (OuterVolumeSpecName: "certs") pod "a17438b6-e92b-4e6e-90dd-6843714adc35" (UID: "a17438b6-e92b-4e6e-90dd-6843714adc35"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.458039 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a17438b6-e92b-4e6e-90dd-6843714adc35-logs" (OuterVolumeSpecName: "logs") pod "a17438b6-e92b-4e6e-90dd-6843714adc35" (UID: "a17438b6-e92b-4e6e-90dd-6843714adc35"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.479377 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-config-data" (OuterVolumeSpecName: "config-data") pod "a17438b6-e92b-4e6e-90dd-6843714adc35" (UID: "a17438b6-e92b-4e6e-90dd-6843714adc35"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.490234 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a17438b6-e92b-4e6e-90dd-6843714adc35" (UID: "a17438b6-e92b-4e6e-90dd-6843714adc35"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.551164 4696 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/a17438b6-e92b-4e6e-90dd-6843714adc35-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.551197 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.551207 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.551218 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkvfm\" (UniqueName: \"kubernetes.io/projected/a17438b6-e92b-4e6e-90dd-6843714adc35-kube-api-access-nkvfm\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.551226 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.551235 4696 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a17438b6-e92b-4e6e-90dd-6843714adc35-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.551243 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a17438b6-e92b-4e6e-90dd-6843714adc35-logs\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.573782 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19b4b8b0-047b-48bf-9c9b-70fe162cf5bd" path="/var/lib/kubelet/pods/19b4b8b0-047b-48bf-9c9b-70fe162cf5bd/volumes" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.578225 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="235bf627-f528-4aa9-85c1-b069ac61f695" path="/var/lib/kubelet/pods/235bf627-f528-4aa9-85c1-b069ac61f695/volumes" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.579332 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbd625f3-5fcc-48f5-9022-df08fdc01887" path="/var/lib/kubelet/pods/dbd625f3-5fcc-48f5-9022-df08fdc01887/volumes" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.580368 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e341e7e5-528c-42cb-b3df-725843538f37" path="/var/lib/kubelet/pods/e341e7e5-528c-42cb-b3df-725843538f37/volumes" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.688220 4696 generic.go:334] "Generic (PLEG): container finished" podID="a17438b6-e92b-4e6e-90dd-6843714adc35" containerID="3f8919fe290a02f8091f7acbae0079823adb80b458b523abc25f4c7d3b864940" exitCode=0 Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.688555 4696 generic.go:334] "Generic (PLEG): container finished" podID="a17438b6-e92b-4e6e-90dd-6843714adc35" containerID="18367b044502afe3de0a6c99ba84b504c44a31dc6a00ff4630ce9973c48c5660" exitCode=143 Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.688563 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"a17438b6-e92b-4e6e-90dd-6843714adc35","Type":"ContainerDied","Data":"3f8919fe290a02f8091f7acbae0079823adb80b458b523abc25f4c7d3b864940"} Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.688631 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"a17438b6-e92b-4e6e-90dd-6843714adc35","Type":"ContainerDied","Data":"18367b044502afe3de0a6c99ba84b504c44a31dc6a00ff4630ce9973c48c5660"} Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.688652 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.688649 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"a17438b6-e92b-4e6e-90dd-6843714adc35","Type":"ContainerDied","Data":"eb4aa5abef4754dcdae58575b39384b5119cd9cd8be3b7691349867e4a4d4718"} Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.688699 4696 scope.go:117] "RemoveContainer" containerID="3f8919fe290a02f8091f7acbae0079823adb80b458b523abc25f4c7d3b864940" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.700192 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8d228678-e5cf-4da4-9a96-7ed39576c142","Type":"ContainerStarted","Data":"cc0ff885d9078ca9dbd72b8d4086da743400cbb01ef9530975ee469345e849ad"} Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.703060 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-proc-0" podUID="ba009f97-23a1-4296-8f34-aeb08fcc6b80" containerName="cloudkitty-proc" containerID="cri-o://e563f58ed0293b5674d023311cc2edf051d98f1f64074f381b1c742fad20ec28" gracePeriod=30 Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.754757 4696 scope.go:117] "RemoveContainer" containerID="18367b044502afe3de0a6c99ba84b504c44a31dc6a00ff4630ce9973c48c5660" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.762074 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.780291 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-api-0"] Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.826996 4696 scope.go:117] "RemoveContainer" containerID="3f8919fe290a02f8091f7acbae0079823adb80b458b523abc25f4c7d3b864940" Mar 21 08:48:16 crc kubenswrapper[4696]: E0321 08:48:16.831323 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f8919fe290a02f8091f7acbae0079823adb80b458b523abc25f4c7d3b864940\": container with ID starting with 3f8919fe290a02f8091f7acbae0079823adb80b458b523abc25f4c7d3b864940 not found: ID does not exist" containerID="3f8919fe290a02f8091f7acbae0079823adb80b458b523abc25f4c7d3b864940" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.831364 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f8919fe290a02f8091f7acbae0079823adb80b458b523abc25f4c7d3b864940"} err="failed to get container status \"3f8919fe290a02f8091f7acbae0079823adb80b458b523abc25f4c7d3b864940\": rpc error: code = NotFound desc = could not find container \"3f8919fe290a02f8091f7acbae0079823adb80b458b523abc25f4c7d3b864940\": container with ID starting with 3f8919fe290a02f8091f7acbae0079823adb80b458b523abc25f4c7d3b864940 not found: ID does not exist" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.831391 4696 scope.go:117] "RemoveContainer" containerID="18367b044502afe3de0a6c99ba84b504c44a31dc6a00ff4630ce9973c48c5660" Mar 21 08:48:16 crc kubenswrapper[4696]: E0321 08:48:16.831594 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18367b044502afe3de0a6c99ba84b504c44a31dc6a00ff4630ce9973c48c5660\": container with ID starting with 18367b044502afe3de0a6c99ba84b504c44a31dc6a00ff4630ce9973c48c5660 not found: ID does not exist" containerID="18367b044502afe3de0a6c99ba84b504c44a31dc6a00ff4630ce9973c48c5660" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.831620 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18367b044502afe3de0a6c99ba84b504c44a31dc6a00ff4630ce9973c48c5660"} err="failed to get container status \"18367b044502afe3de0a6c99ba84b504c44a31dc6a00ff4630ce9973c48c5660\": rpc error: code = NotFound desc = could not find container \"18367b044502afe3de0a6c99ba84b504c44a31dc6a00ff4630ce9973c48c5660\": container with ID starting with 18367b044502afe3de0a6c99ba84b504c44a31dc6a00ff4630ce9973c48c5660 not found: ID does not exist" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.831634 4696 scope.go:117] "RemoveContainer" containerID="3f8919fe290a02f8091f7acbae0079823adb80b458b523abc25f4c7d3b864940" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.831795 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f8919fe290a02f8091f7acbae0079823adb80b458b523abc25f4c7d3b864940"} err="failed to get container status \"3f8919fe290a02f8091f7acbae0079823adb80b458b523abc25f4c7d3b864940\": rpc error: code = NotFound desc = could not find container \"3f8919fe290a02f8091f7acbae0079823adb80b458b523abc25f4c7d3b864940\": container with ID starting with 3f8919fe290a02f8091f7acbae0079823adb80b458b523abc25f4c7d3b864940 not found: ID does not exist" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.831888 4696 scope.go:117] "RemoveContainer" containerID="18367b044502afe3de0a6c99ba84b504c44a31dc6a00ff4630ce9973c48c5660" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.832059 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18367b044502afe3de0a6c99ba84b504c44a31dc6a00ff4630ce9973c48c5660"} err="failed to get container status \"18367b044502afe3de0a6c99ba84b504c44a31dc6a00ff4630ce9973c48c5660\": rpc error: code = NotFound desc = could not find container \"18367b044502afe3de0a6c99ba84b504c44a31dc6a00ff4630ce9973c48c5660\": container with ID starting with 18367b044502afe3de0a6c99ba84b504c44a31dc6a00ff4630ce9973c48c5660 not found: ID does not exist" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.847636 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-api-0"] Mar 21 08:48:16 crc kubenswrapper[4696]: E0321 08:48:16.848124 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a17438b6-e92b-4e6e-90dd-6843714adc35" containerName="cloudkitty-api" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.848141 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="a17438b6-e92b-4e6e-90dd-6843714adc35" containerName="cloudkitty-api" Mar 21 08:48:16 crc kubenswrapper[4696]: E0321 08:48:16.848181 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a17438b6-e92b-4e6e-90dd-6843714adc35" containerName="cloudkitty-api-log" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.848188 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="a17438b6-e92b-4e6e-90dd-6843714adc35" containerName="cloudkitty-api-log" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.848386 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="a17438b6-e92b-4e6e-90dd-6843714adc35" containerName="cloudkitty-api" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.848404 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="a17438b6-e92b-4e6e-90dd-6843714adc35" containerName="cloudkitty-api-log" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.849921 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.855112 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-api-config-data" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.855254 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-internal-svc" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.855394 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-public-svc" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.934910 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.966857 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.966900 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.966951 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.967005 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crcb9\" (UniqueName: \"kubernetes.io/projected/d623e522-1d67-45ef-8ebd-a8a37e12f045-kube-api-access-crcb9\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.967029 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.967081 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d623e522-1d67-45ef-8ebd-a8a37e12f045-logs\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.967104 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d623e522-1d67-45ef-8ebd-a8a37e12f045-certs\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.967128 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-scripts\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:16 crc kubenswrapper[4696]: I0321 08:48:16.967145 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-config-data\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.076003 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.076215 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.076263 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.076309 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crcb9\" (UniqueName: \"kubernetes.io/projected/d623e522-1d67-45ef-8ebd-a8a37e12f045-kube-api-access-crcb9\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.076335 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.076386 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d623e522-1d67-45ef-8ebd-a8a37e12f045-logs\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.076413 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d623e522-1d67-45ef-8ebd-a8a37e12f045-certs\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.076434 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-scripts\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.076448 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-config-data\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.078249 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d623e522-1d67-45ef-8ebd-a8a37e12f045-logs\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.090542 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-scripts\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.091785 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.092452 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.093496 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d623e522-1d67-45ef-8ebd-a8a37e12f045-certs\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.100940 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.101257 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.101856 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-config-data\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.103982 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.111309 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crcb9\" (UniqueName: \"kubernetes.io/projected/d623e522-1d67-45ef-8ebd-a8a37e12f045-kube-api-access-crcb9\") pod \"cloudkitty-api-0\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " pod="openstack/cloudkitty-api-0" Mar 21 08:48:17 crc kubenswrapper[4696]: W0321 08:48:17.134226 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e007714_252e_43e9_b281_323db9d580ad.slice/crio-27d84e25c5928559268ad7a13b8eba6d0a7c02093f825ef16ae5fa6da1d8e909 WatchSource:0}: Error finding container 27d84e25c5928559268ad7a13b8eba6d0a7c02093f825ef16ae5fa6da1d8e909: Status 404 returned error can't find the container with id 27d84e25c5928559268ad7a13b8eba6d0a7c02093f825ef16ae5fa6da1d8e909 Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.322867 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.802049 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8d228678-e5cf-4da4-9a96-7ed39576c142","Type":"ContainerStarted","Data":"9d9dfc48120f42227240e5760169bfefe2c7cc423a8a2f2b188281f8a733aa4f"} Mar 21 08:48:17 crc kubenswrapper[4696]: I0321 08:48:17.812467 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e007714-252e-43e9-b281-323db9d580ad","Type":"ContainerStarted","Data":"27d84e25c5928559268ad7a13b8eba6d0a7c02093f825ef16ae5fa6da1d8e909"} Mar 21 08:48:18 crc kubenswrapper[4696]: I0321 08:48:18.141791 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Mar 21 08:48:18 crc kubenswrapper[4696]: I0321 08:48:18.584372 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a17438b6-e92b-4e6e-90dd-6843714adc35" path="/var/lib/kubelet/pods/a17438b6-e92b-4e6e-90dd-6843714adc35/volumes" Mar 21 08:48:18 crc kubenswrapper[4696]: I0321 08:48:18.829705 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"d623e522-1d67-45ef-8ebd-a8a37e12f045","Type":"ContainerStarted","Data":"122f960a9aa1d8bd0345a90c369c67c7c94fa1b94e8fcf6019667927ee380e6a"} Mar 21 08:48:18 crc kubenswrapper[4696]: I0321 08:48:18.829747 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"d623e522-1d67-45ef-8ebd-a8a37e12f045","Type":"ContainerStarted","Data":"d55f509a352fa8a26feffdf52136d37fceac48489792434ec6dbf155bca37171"} Mar 21 08:48:18 crc kubenswrapper[4696]: I0321 08:48:18.841072 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8d228678-e5cf-4da4-9a96-7ed39576c142","Type":"ContainerStarted","Data":"00e191dc3e3af2d8b7543bf06a5d5bc7dab9544bcc93be7d8fd235c8365dcbda"} Mar 21 08:48:18 crc kubenswrapper[4696]: I0321 08:48:18.841440 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 21 08:48:18 crc kubenswrapper[4696]: I0321 08:48:18.847157 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e007714-252e-43e9-b281-323db9d580ad","Type":"ContainerStarted","Data":"fa78034c74c9f61a2e6cff2f20c9d14454749e1e7b00d112e2dadbe48a3560cb"} Mar 21 08:48:18 crc kubenswrapper[4696]: I0321 08:48:18.869461 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.8694341439999995 podStartE2EDuration="4.869434144s" podCreationTimestamp="2026-03-21 08:48:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:48:18.859919681 +0000 UTC m=+1232.980800404" watchObservedRunningTime="2026-03-21 08:48:18.869434144 +0000 UTC m=+1232.990314857" Mar 21 08:48:18 crc kubenswrapper[4696]: I0321 08:48:18.910967 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 21 08:48:19 crc kubenswrapper[4696]: I0321 08:48:19.266144 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 21 08:48:19 crc kubenswrapper[4696]: I0321 08:48:19.858122 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e007714-252e-43e9-b281-323db9d580ad","Type":"ContainerStarted","Data":"1ec234430915a7754af8a24e7b6a89c1a27fbe794f01a4f8c0b2b4cc619c41bf"} Mar 21 08:48:19 crc kubenswrapper[4696]: I0321 08:48:19.860570 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"d623e522-1d67-45ef-8ebd-a8a37e12f045","Type":"ContainerStarted","Data":"68c5a483e2f9a4917af42f3a50fcac8a07ccc00aa1a9b31b189b930f3f5d071c"} Mar 21 08:48:19 crc kubenswrapper[4696]: I0321 08:48:19.860927 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-api-0" Mar 21 08:48:19 crc kubenswrapper[4696]: I0321 08:48:19.862882 4696 generic.go:334] "Generic (PLEG): container finished" podID="ba009f97-23a1-4296-8f34-aeb08fcc6b80" containerID="e563f58ed0293b5674d023311cc2edf051d98f1f64074f381b1c742fad20ec28" exitCode=0 Mar 21 08:48:19 crc kubenswrapper[4696]: I0321 08:48:19.862956 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"ba009f97-23a1-4296-8f34-aeb08fcc6b80","Type":"ContainerDied","Data":"e563f58ed0293b5674d023311cc2edf051d98f1f64074f381b1c742fad20ec28"} Mar 21 08:48:19 crc kubenswrapper[4696]: I0321 08:48:19.889617 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-api-0" podStartSLOduration=3.889601544 podStartE2EDuration="3.889601544s" podCreationTimestamp="2026-03-21 08:48:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:48:19.885883502 +0000 UTC m=+1234.006764215" watchObservedRunningTime="2026-03-21 08:48:19.889601544 +0000 UTC m=+1234.010482247" Mar 21 08:48:19 crc kubenswrapper[4696]: I0321 08:48:19.915496 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Mar 21 08:48:19 crc kubenswrapper[4696]: I0321 08:48:19.945874 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.047301 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-config-data-custom\") pod \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.047383 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-config-data\") pod \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.047427 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/ba009f97-23a1-4296-8f34-aeb08fcc6b80-certs\") pod \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.047451 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwrmx\" (UniqueName: \"kubernetes.io/projected/ba009f97-23a1-4296-8f34-aeb08fcc6b80-kube-api-access-pwrmx\") pod \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.047530 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-scripts\") pod \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.047640 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-combined-ca-bundle\") pod \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\" (UID: \"ba009f97-23a1-4296-8f34-aeb08fcc6b80\") " Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.053239 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba009f97-23a1-4296-8f34-aeb08fcc6b80-kube-api-access-pwrmx" (OuterVolumeSpecName: "kube-api-access-pwrmx") pod "ba009f97-23a1-4296-8f34-aeb08fcc6b80" (UID: "ba009f97-23a1-4296-8f34-aeb08fcc6b80"). InnerVolumeSpecName "kube-api-access-pwrmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.055555 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-scripts" (OuterVolumeSpecName: "scripts") pod "ba009f97-23a1-4296-8f34-aeb08fcc6b80" (UID: "ba009f97-23a1-4296-8f34-aeb08fcc6b80"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.056671 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ba009f97-23a1-4296-8f34-aeb08fcc6b80" (UID: "ba009f97-23a1-4296-8f34-aeb08fcc6b80"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.066050 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba009f97-23a1-4296-8f34-aeb08fcc6b80-certs" (OuterVolumeSpecName: "certs") pod "ba009f97-23a1-4296-8f34-aeb08fcc6b80" (UID: "ba009f97-23a1-4296-8f34-aeb08fcc6b80"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.088902 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ba009f97-23a1-4296-8f34-aeb08fcc6b80" (UID: "ba009f97-23a1-4296-8f34-aeb08fcc6b80"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.103404 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-config-data" (OuterVolumeSpecName: "config-data") pod "ba009f97-23a1-4296-8f34-aeb08fcc6b80" (UID: "ba009f97-23a1-4296-8f34-aeb08fcc6b80"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.150003 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.150043 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.150056 4696 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.150064 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba009f97-23a1-4296-8f34-aeb08fcc6b80-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.150072 4696 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/ba009f97-23a1-4296-8f34-aeb08fcc6b80-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.150082 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwrmx\" (UniqueName: \"kubernetes.io/projected/ba009f97-23a1-4296-8f34-aeb08fcc6b80-kube-api-access-pwrmx\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.339713 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-696f547b5d-r9gcd" Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.499054 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.563475 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-6jw7g"] Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.563700 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" podUID="c0e814b1-d0f8-4a97-a76e-da97f375c9a9" containerName="dnsmasq-dns" containerID="cri-o://d7e8f9fab679381fe54039871691160d657f02791447d84979fe4b5b47767dc0" gracePeriod=10 Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.886862 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"ba009f97-23a1-4296-8f34-aeb08fcc6b80","Type":"ContainerDied","Data":"0ce2ebece7506499e7f3d5dc03ee27110d15a1383a50dd10f282d5cab8967d45"} Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.887161 4696 scope.go:117] "RemoveContainer" containerID="e563f58ed0293b5674d023311cc2edf051d98f1f64074f381b1c742fad20ec28" Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.887054 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.895968 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e007714-252e-43e9-b281-323db9d580ad","Type":"ContainerStarted","Data":"0f683af8454805602847c08cf6ff20f6a5a145e396e630d81e262794d191a866"} Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.897845 4696 generic.go:334] "Generic (PLEG): container finished" podID="c0e814b1-d0f8-4a97-a76e-da97f375c9a9" containerID="d7e8f9fab679381fe54039871691160d657f02791447d84979fe4b5b47767dc0" exitCode=0 Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.898024 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="33133b7f-8c96-4813-b1ab-2a40acb2cd84" containerName="cinder-scheduler" containerID="cri-o://cc664ad69a991d8883bbae1c6ec054b7b3b665f8bd2f4a92612ba533f977da0f" gracePeriod=30 Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.898195 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" event={"ID":"c0e814b1-d0f8-4a97-a76e-da97f375c9a9","Type":"ContainerDied","Data":"d7e8f9fab679381fe54039871691160d657f02791447d84979fe4b5b47767dc0"} Mar 21 08:48:20 crc kubenswrapper[4696]: I0321 08:48:20.900081 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="33133b7f-8c96-4813-b1ab-2a40acb2cd84" containerName="probe" containerID="cri-o://db14586774b84c9cfaa3cc665fa911d8ac8c18592c7f5dac9340bfddddc13b20" gracePeriod=30 Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.001153 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.022421 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-proc-0"] Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.031168 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-proc-0"] Mar 21 08:48:21 crc kubenswrapper[4696]: E0321 08:48:21.031783 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba009f97-23a1-4296-8f34-aeb08fcc6b80" containerName="cloudkitty-proc" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.031800 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba009f97-23a1-4296-8f34-aeb08fcc6b80" containerName="cloudkitty-proc" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.032008 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba009f97-23a1-4296-8f34-aeb08fcc6b80" containerName="cloudkitty-proc" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.032890 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.037160 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-proc-config-data" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.057962 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.090267 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/623a2b50-8a77-47df-800c-5ec7c237255c-certs\") pod \"cloudkitty-proc-0\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.090555 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.090610 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.090636 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs2sr\" (UniqueName: \"kubernetes.io/projected/623a2b50-8a77-47df-800c-5ec7c237255c-kube-api-access-fs2sr\") pod \"cloudkitty-proc-0\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.090892 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-scripts\") pod \"cloudkitty-proc-0\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.090934 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-config-data\") pod \"cloudkitty-proc-0\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.193663 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/623a2b50-8a77-47df-800c-5ec7c237255c-certs\") pod \"cloudkitty-proc-0\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.193726 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.193757 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.193779 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs2sr\" (UniqueName: \"kubernetes.io/projected/623a2b50-8a77-47df-800c-5ec7c237255c-kube-api-access-fs2sr\") pod \"cloudkitty-proc-0\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.193869 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-scripts\") pod \"cloudkitty-proc-0\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.193887 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-config-data\") pod \"cloudkitty-proc-0\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.199597 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-scripts\") pod \"cloudkitty-proc-0\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.201627 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.202690 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-config-data\") pod \"cloudkitty-proc-0\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.204342 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/623a2b50-8a77-47df-800c-5ec7c237255c-certs\") pod \"cloudkitty-proc-0\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.212986 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.218478 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs2sr\" (UniqueName: \"kubernetes.io/projected/623a2b50-8a77-47df-800c-5ec7c237255c-kube-api-access-fs2sr\") pod \"cloudkitty-proc-0\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.298771 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.355629 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.397203 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mmx2\" (UniqueName: \"kubernetes.io/projected/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-kube-api-access-2mmx2\") pod \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.397322 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-ovsdbserver-nb\") pod \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.397384 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-dns-svc\") pod \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.397526 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-ovsdbserver-sb\") pod \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.397549 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-config\") pod \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.397566 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-dns-swift-storage-0\") pod \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\" (UID: \"c0e814b1-d0f8-4a97-a76e-da97f375c9a9\") " Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.418083 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-kube-api-access-2mmx2" (OuterVolumeSpecName: "kube-api-access-2mmx2") pod "c0e814b1-d0f8-4a97-a76e-da97f375c9a9" (UID: "c0e814b1-d0f8-4a97-a76e-da97f375c9a9"). InnerVolumeSpecName "kube-api-access-2mmx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.500258 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mmx2\" (UniqueName: \"kubernetes.io/projected/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-kube-api-access-2mmx2\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.560972 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-config" (OuterVolumeSpecName: "config") pod "c0e814b1-d0f8-4a97-a76e-da97f375c9a9" (UID: "c0e814b1-d0f8-4a97-a76e-da97f375c9a9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.573325 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c0e814b1-d0f8-4a97-a76e-da97f375c9a9" (UID: "c0e814b1-d0f8-4a97-a76e-da97f375c9a9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.607557 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c0e814b1-d0f8-4a97-a76e-da97f375c9a9" (UID: "c0e814b1-d0f8-4a97-a76e-da97f375c9a9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.609476 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.609509 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.609522 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.625350 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c0e814b1-d0f8-4a97-a76e-da97f375c9a9" (UID: "c0e814b1-d0f8-4a97-a76e-da97f375c9a9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.711686 4696 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.742393 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c0e814b1-d0f8-4a97-a76e-da97f375c9a9" (UID: "c0e814b1-d0f8-4a97-a76e-da97f375c9a9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.813319 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c0e814b1-d0f8-4a97-a76e-da97f375c9a9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.928516 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" event={"ID":"c0e814b1-d0f8-4a97-a76e-da97f375c9a9","Type":"ContainerDied","Data":"5153db0aeb397e48b4ed0cffa78cfe6af26091044bf749685d6054821e78a5d9"} Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.928775 4696 scope.go:117] "RemoveContainer" containerID="d7e8f9fab679381fe54039871691160d657f02791447d84979fe4b5b47767dc0" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.928641 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-6jw7g" Mar 21 08:48:21 crc kubenswrapper[4696]: I0321 08:48:21.975506 4696 scope.go:117] "RemoveContainer" containerID="bf696af1ec291a1084ebc174abfdd15a2b74d5894459e8b56e1bc8835cac1e17" Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.007516 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-6jw7g"] Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.021459 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-6jw7g"] Mar 21 08:48:22 crc kubenswrapper[4696]: W0321 08:48:22.086023 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod623a2b50_8a77_47df_800c_5ec7c237255c.slice/crio-7368dbc952d8a83cdc4e59b462d771cda60acfd48762b5f9422ce41553ed3d96 WatchSource:0}: Error finding container 7368dbc952d8a83cdc4e59b462d771cda60acfd48762b5f9422ce41553ed3d96: Status 404 returned error can't find the container with id 7368dbc952d8a83cdc4e59b462d771cda60acfd48762b5f9422ce41553ed3d96 Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.086190 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.547808 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba009f97-23a1-4296-8f34-aeb08fcc6b80" path="/var/lib/kubelet/pods/ba009f97-23a1-4296-8f34-aeb08fcc6b80/volumes" Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.549212 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0e814b1-d0f8-4a97-a76e-da97f375c9a9" path="/var/lib/kubelet/pods/c0e814b1-d0f8-4a97-a76e-da97f375c9a9/volumes" Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.642384 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.736431 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/33133b7f-8c96-4813-b1ab-2a40acb2cd84-etc-machine-id\") pod \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.736487 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hb7kr\" (UniqueName: \"kubernetes.io/projected/33133b7f-8c96-4813-b1ab-2a40acb2cd84-kube-api-access-hb7kr\") pod \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.736510 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/33133b7f-8c96-4813-b1ab-2a40acb2cd84-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "33133b7f-8c96-4813-b1ab-2a40acb2cd84" (UID: "33133b7f-8c96-4813-b1ab-2a40acb2cd84"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.736630 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-scripts\") pod \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.736652 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-combined-ca-bundle\") pod \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.736681 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-config-data\") pod \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.736714 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-config-data-custom\") pod \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\" (UID: \"33133b7f-8c96-4813-b1ab-2a40acb2cd84\") " Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.737192 4696 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/33133b7f-8c96-4813-b1ab-2a40acb2cd84-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.741987 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "33133b7f-8c96-4813-b1ab-2a40acb2cd84" (UID: "33133b7f-8c96-4813-b1ab-2a40acb2cd84"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.742085 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-scripts" (OuterVolumeSpecName: "scripts") pod "33133b7f-8c96-4813-b1ab-2a40acb2cd84" (UID: "33133b7f-8c96-4813-b1ab-2a40acb2cd84"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.763133 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33133b7f-8c96-4813-b1ab-2a40acb2cd84-kube-api-access-hb7kr" (OuterVolumeSpecName: "kube-api-access-hb7kr") pod "33133b7f-8c96-4813-b1ab-2a40acb2cd84" (UID: "33133b7f-8c96-4813-b1ab-2a40acb2cd84"). InnerVolumeSpecName "kube-api-access-hb7kr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.816128 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33133b7f-8c96-4813-b1ab-2a40acb2cd84" (UID: "33133b7f-8c96-4813-b1ab-2a40acb2cd84"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.839366 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.839403 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.839417 4696 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.839426 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hb7kr\" (UniqueName: \"kubernetes.io/projected/33133b7f-8c96-4813-b1ab-2a40acb2cd84-kube-api-access-hb7kr\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.862696 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-config-data" (OuterVolumeSpecName: "config-data") pod "33133b7f-8c96-4813-b1ab-2a40acb2cd84" (UID: "33133b7f-8c96-4813-b1ab-2a40acb2cd84"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.939420 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e007714-252e-43e9-b281-323db9d580ad","Type":"ContainerStarted","Data":"672e68b1de41b8ea639fbd69f58f1e6062fbf25d20341f77c67fd93a4b9c35a7"} Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.940698 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33133b7f-8c96-4813-b1ab-2a40acb2cd84-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.940970 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.944086 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"623a2b50-8a77-47df-800c-5ec7c237255c","Type":"ContainerStarted","Data":"b1600882e3e31cebfdf37e53812fcf328b05af85029dd298a9919826dcf85349"} Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.944122 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"623a2b50-8a77-47df-800c-5ec7c237255c","Type":"ContainerStarted","Data":"7368dbc952d8a83cdc4e59b462d771cda60acfd48762b5f9422ce41553ed3d96"} Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.945864 4696 generic.go:334] "Generic (PLEG): container finished" podID="33133b7f-8c96-4813-b1ab-2a40acb2cd84" containerID="db14586774b84c9cfaa3cc665fa911d8ac8c18592c7f5dac9340bfddddc13b20" exitCode=0 Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.945888 4696 generic.go:334] "Generic (PLEG): container finished" podID="33133b7f-8c96-4813-b1ab-2a40acb2cd84" containerID="cc664ad69a991d8883bbae1c6ec054b7b3b665f8bd2f4a92612ba533f977da0f" exitCode=0 Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.945893 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"33133b7f-8c96-4813-b1ab-2a40acb2cd84","Type":"ContainerDied","Data":"db14586774b84c9cfaa3cc665fa911d8ac8c18592c7f5dac9340bfddddc13b20"} Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.945926 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"33133b7f-8c96-4813-b1ab-2a40acb2cd84","Type":"ContainerDied","Data":"cc664ad69a991d8883bbae1c6ec054b7b3b665f8bd2f4a92612ba533f977da0f"} Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.945938 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"33133b7f-8c96-4813-b1ab-2a40acb2cd84","Type":"ContainerDied","Data":"5a2aa71528c016a1e6faab82e79b728928036f78624c31bdce4e952c9b4bf5f0"} Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.945922 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.945986 4696 scope.go:117] "RemoveContainer" containerID="db14586774b84c9cfaa3cc665fa911d8ac8c18592c7f5dac9340bfddddc13b20" Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.974111 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.538950191 podStartE2EDuration="7.974088448s" podCreationTimestamp="2026-03-21 08:48:15 +0000 UTC" firstStartedPulling="2026-03-21 08:48:17.141777752 +0000 UTC m=+1231.262658465" lastFinishedPulling="2026-03-21 08:48:21.576916009 +0000 UTC m=+1235.697796722" observedRunningTime="2026-03-21 08:48:22.96114855 +0000 UTC m=+1237.082029263" watchObservedRunningTime="2026-03-21 08:48:22.974088448 +0000 UTC m=+1237.094969161" Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.984614 4696 scope.go:117] "RemoveContainer" containerID="cc664ad69a991d8883bbae1c6ec054b7b3b665f8bd2f4a92612ba533f977da0f" Mar 21 08:48:22 crc kubenswrapper[4696]: I0321 08:48:22.996020 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.021495 4696 scope.go:117] "RemoveContainer" containerID="db14586774b84c9cfaa3cc665fa911d8ac8c18592c7f5dac9340bfddddc13b20" Mar 21 08:48:23 crc kubenswrapper[4696]: E0321 08:48:23.023154 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db14586774b84c9cfaa3cc665fa911d8ac8c18592c7f5dac9340bfddddc13b20\": container with ID starting with db14586774b84c9cfaa3cc665fa911d8ac8c18592c7f5dac9340bfddddc13b20 not found: ID does not exist" containerID="db14586774b84c9cfaa3cc665fa911d8ac8c18592c7f5dac9340bfddddc13b20" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.023193 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db14586774b84c9cfaa3cc665fa911d8ac8c18592c7f5dac9340bfddddc13b20"} err="failed to get container status \"db14586774b84c9cfaa3cc665fa911d8ac8c18592c7f5dac9340bfddddc13b20\": rpc error: code = NotFound desc = could not find container \"db14586774b84c9cfaa3cc665fa911d8ac8c18592c7f5dac9340bfddddc13b20\": container with ID starting with db14586774b84c9cfaa3cc665fa911d8ac8c18592c7f5dac9340bfddddc13b20 not found: ID does not exist" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.023218 4696 scope.go:117] "RemoveContainer" containerID="cc664ad69a991d8883bbae1c6ec054b7b3b665f8bd2f4a92612ba533f977da0f" Mar 21 08:48:23 crc kubenswrapper[4696]: E0321 08:48:23.023421 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc664ad69a991d8883bbae1c6ec054b7b3b665f8bd2f4a92612ba533f977da0f\": container with ID starting with cc664ad69a991d8883bbae1c6ec054b7b3b665f8bd2f4a92612ba533f977da0f not found: ID does not exist" containerID="cc664ad69a991d8883bbae1c6ec054b7b3b665f8bd2f4a92612ba533f977da0f" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.023444 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc664ad69a991d8883bbae1c6ec054b7b3b665f8bd2f4a92612ba533f977da0f"} err="failed to get container status \"cc664ad69a991d8883bbae1c6ec054b7b3b665f8bd2f4a92612ba533f977da0f\": rpc error: code = NotFound desc = could not find container \"cc664ad69a991d8883bbae1c6ec054b7b3b665f8bd2f4a92612ba533f977da0f\": container with ID starting with cc664ad69a991d8883bbae1c6ec054b7b3b665f8bd2f4a92612ba533f977da0f not found: ID does not exist" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.023471 4696 scope.go:117] "RemoveContainer" containerID="db14586774b84c9cfaa3cc665fa911d8ac8c18592c7f5dac9340bfddddc13b20" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.023634 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db14586774b84c9cfaa3cc665fa911d8ac8c18592c7f5dac9340bfddddc13b20"} err="failed to get container status \"db14586774b84c9cfaa3cc665fa911d8ac8c18592c7f5dac9340bfddddc13b20\": rpc error: code = NotFound desc = could not find container \"db14586774b84c9cfaa3cc665fa911d8ac8c18592c7f5dac9340bfddddc13b20\": container with ID starting with db14586774b84c9cfaa3cc665fa911d8ac8c18592c7f5dac9340bfddddc13b20 not found: ID does not exist" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.023674 4696 scope.go:117] "RemoveContainer" containerID="cc664ad69a991d8883bbae1c6ec054b7b3b665f8bd2f4a92612ba533f977da0f" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.023851 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc664ad69a991d8883bbae1c6ec054b7b3b665f8bd2f4a92612ba533f977da0f"} err="failed to get container status \"cc664ad69a991d8883bbae1c6ec054b7b3b665f8bd2f4a92612ba533f977da0f\": rpc error: code = NotFound desc = could not find container \"cc664ad69a991d8883bbae1c6ec054b7b3b665f8bd2f4a92612ba533f977da0f\": container with ID starting with cc664ad69a991d8883bbae1c6ec054b7b3b665f8bd2f4a92612ba533f977da0f not found: ID does not exist" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.028888 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.049212 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 21 08:48:23 crc kubenswrapper[4696]: E0321 08:48:23.050002 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0e814b1-d0f8-4a97-a76e-da97f375c9a9" containerName="dnsmasq-dns" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.050125 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0e814b1-d0f8-4a97-a76e-da97f375c9a9" containerName="dnsmasq-dns" Mar 21 08:48:23 crc kubenswrapper[4696]: E0321 08:48:23.050209 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33133b7f-8c96-4813-b1ab-2a40acb2cd84" containerName="cinder-scheduler" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.050288 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="33133b7f-8c96-4813-b1ab-2a40acb2cd84" containerName="cinder-scheduler" Mar 21 08:48:23 crc kubenswrapper[4696]: E0321 08:48:23.050369 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0e814b1-d0f8-4a97-a76e-da97f375c9a9" containerName="init" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.050437 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0e814b1-d0f8-4a97-a76e-da97f375c9a9" containerName="init" Mar 21 08:48:23 crc kubenswrapper[4696]: E0321 08:48:23.050540 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33133b7f-8c96-4813-b1ab-2a40acb2cd84" containerName="probe" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.050623 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="33133b7f-8c96-4813-b1ab-2a40acb2cd84" containerName="probe" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.050956 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="33133b7f-8c96-4813-b1ab-2a40acb2cd84" containerName="probe" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.051091 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="33133b7f-8c96-4813-b1ab-2a40acb2cd84" containerName="cinder-scheduler" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.051176 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0e814b1-d0f8-4a97-a76e-da97f375c9a9" containerName="dnsmasq-dns" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.052877 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.057956 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-proc-0" podStartSLOduration=3.05793644 podStartE2EDuration="3.05793644s" podCreationTimestamp="2026-03-21 08:48:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:48:23.004383578 +0000 UTC m=+1237.125264291" watchObservedRunningTime="2026-03-21 08:48:23.05793644 +0000 UTC m=+1237.178817153" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.081205 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.101873 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.144919 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cs47w\" (UniqueName: \"kubernetes.io/projected/4fce4b85-2f39-4702-b221-fd5e870e6a2d-kube-api-access-cs47w\") pod \"cinder-scheduler-0\" (UID: \"4fce4b85-2f39-4702-b221-fd5e870e6a2d\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.144984 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4fce4b85-2f39-4702-b221-fd5e870e6a2d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4fce4b85-2f39-4702-b221-fd5e870e6a2d\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.145012 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4fce4b85-2f39-4702-b221-fd5e870e6a2d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4fce4b85-2f39-4702-b221-fd5e870e6a2d\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.145036 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fce4b85-2f39-4702-b221-fd5e870e6a2d-scripts\") pod \"cinder-scheduler-0\" (UID: \"4fce4b85-2f39-4702-b221-fd5e870e6a2d\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.145146 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fce4b85-2f39-4702-b221-fd5e870e6a2d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4fce4b85-2f39-4702-b221-fd5e870e6a2d\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.145195 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fce4b85-2f39-4702-b221-fd5e870e6a2d-config-data\") pod \"cinder-scheduler-0\" (UID: \"4fce4b85-2f39-4702-b221-fd5e870e6a2d\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.246530 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fce4b85-2f39-4702-b221-fd5e870e6a2d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4fce4b85-2f39-4702-b221-fd5e870e6a2d\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.246949 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fce4b85-2f39-4702-b221-fd5e870e6a2d-config-data\") pod \"cinder-scheduler-0\" (UID: \"4fce4b85-2f39-4702-b221-fd5e870e6a2d\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.247101 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cs47w\" (UniqueName: \"kubernetes.io/projected/4fce4b85-2f39-4702-b221-fd5e870e6a2d-kube-api-access-cs47w\") pod \"cinder-scheduler-0\" (UID: \"4fce4b85-2f39-4702-b221-fd5e870e6a2d\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.247159 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4fce4b85-2f39-4702-b221-fd5e870e6a2d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4fce4b85-2f39-4702-b221-fd5e870e6a2d\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.247197 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4fce4b85-2f39-4702-b221-fd5e870e6a2d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4fce4b85-2f39-4702-b221-fd5e870e6a2d\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.247227 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fce4b85-2f39-4702-b221-fd5e870e6a2d-scripts\") pod \"cinder-scheduler-0\" (UID: \"4fce4b85-2f39-4702-b221-fd5e870e6a2d\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.247666 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4fce4b85-2f39-4702-b221-fd5e870e6a2d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4fce4b85-2f39-4702-b221-fd5e870e6a2d\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.251437 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fce4b85-2f39-4702-b221-fd5e870e6a2d-scripts\") pod \"cinder-scheduler-0\" (UID: \"4fce4b85-2f39-4702-b221-fd5e870e6a2d\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.251998 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4fce4b85-2f39-4702-b221-fd5e870e6a2d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4fce4b85-2f39-4702-b221-fd5e870e6a2d\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.252183 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fce4b85-2f39-4702-b221-fd5e870e6a2d-config-data\") pod \"cinder-scheduler-0\" (UID: \"4fce4b85-2f39-4702-b221-fd5e870e6a2d\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.255469 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fce4b85-2f39-4702-b221-fd5e870e6a2d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4fce4b85-2f39-4702-b221-fd5e870e6a2d\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.272135 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cs47w\" (UniqueName: \"kubernetes.io/projected/4fce4b85-2f39-4702-b221-fd5e870e6a2d-kube-api-access-cs47w\") pod \"cinder-scheduler-0\" (UID: \"4fce4b85-2f39-4702-b221-fd5e870e6a2d\") " pod="openstack/cinder-scheduler-0" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.399096 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.918991 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 21 08:48:23 crc kubenswrapper[4696]: W0321 08:48:23.922628 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4fce4b85_2f39_4702_b221_fd5e870e6a2d.slice/crio-892f5700a0346657e34813bf6d601a7b97e831a421f37b80282951f746207a30 WatchSource:0}: Error finding container 892f5700a0346657e34813bf6d601a7b97e831a421f37b80282951f746207a30: Status 404 returned error can't find the container with id 892f5700a0346657e34813bf6d601a7b97e831a421f37b80282951f746207a30 Mar 21 08:48:23 crc kubenswrapper[4696]: I0321 08:48:23.960746 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4fce4b85-2f39-4702-b221-fd5e870e6a2d","Type":"ContainerStarted","Data":"892f5700a0346657e34813bf6d601a7b97e831a421f37b80282951f746207a30"} Mar 21 08:48:24 crc kubenswrapper[4696]: I0321 08:48:24.571552 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33133b7f-8c96-4813-b1ab-2a40acb2cd84" path="/var/lib/kubelet/pods/33133b7f-8c96-4813-b1ab-2a40acb2cd84/volumes" Mar 21 08:48:25 crc kubenswrapper[4696]: I0321 08:48:25.025593 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4fce4b85-2f39-4702-b221-fd5e870e6a2d","Type":"ContainerStarted","Data":"687adfea52c20f7d8ec4771f42321dd88d1c8ac68d0a5c5ed56d7011d668bbfa"} Mar 21 08:48:25 crc kubenswrapper[4696]: I0321 08:48:25.522622 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 21 08:48:25 crc kubenswrapper[4696]: I0321 08:48:25.524087 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 21 08:48:25 crc kubenswrapper[4696]: I0321 08:48:25.528006 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Mar 21 08:48:25 crc kubenswrapper[4696]: I0321 08:48:25.528183 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Mar 21 08:48:25 crc kubenswrapper[4696]: I0321 08:48:25.528928 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-2zts6" Mar 21 08:48:25 crc kubenswrapper[4696]: I0321 08:48:25.538428 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 21 08:48:25 crc kubenswrapper[4696]: I0321 08:48:25.610147 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/383967e9-132f-4355-abbb-5904830137b7-openstack-config-secret\") pod \"openstackclient\" (UID: \"383967e9-132f-4355-abbb-5904830137b7\") " pod="openstack/openstackclient" Mar 21 08:48:25 crc kubenswrapper[4696]: I0321 08:48:25.610244 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/383967e9-132f-4355-abbb-5904830137b7-combined-ca-bundle\") pod \"openstackclient\" (UID: \"383967e9-132f-4355-abbb-5904830137b7\") " pod="openstack/openstackclient" Mar 21 08:48:25 crc kubenswrapper[4696]: I0321 08:48:25.610310 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/383967e9-132f-4355-abbb-5904830137b7-openstack-config\") pod \"openstackclient\" (UID: \"383967e9-132f-4355-abbb-5904830137b7\") " pod="openstack/openstackclient" Mar 21 08:48:25 crc kubenswrapper[4696]: I0321 08:48:25.610371 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntzd6\" (UniqueName: \"kubernetes.io/projected/383967e9-132f-4355-abbb-5904830137b7-kube-api-access-ntzd6\") pod \"openstackclient\" (UID: \"383967e9-132f-4355-abbb-5904830137b7\") " pod="openstack/openstackclient" Mar 21 08:48:25 crc kubenswrapper[4696]: I0321 08:48:25.712334 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/383967e9-132f-4355-abbb-5904830137b7-openstack-config-secret\") pod \"openstackclient\" (UID: \"383967e9-132f-4355-abbb-5904830137b7\") " pod="openstack/openstackclient" Mar 21 08:48:25 crc kubenswrapper[4696]: I0321 08:48:25.712411 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/383967e9-132f-4355-abbb-5904830137b7-combined-ca-bundle\") pod \"openstackclient\" (UID: \"383967e9-132f-4355-abbb-5904830137b7\") " pod="openstack/openstackclient" Mar 21 08:48:25 crc kubenswrapper[4696]: I0321 08:48:25.712464 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/383967e9-132f-4355-abbb-5904830137b7-openstack-config\") pod \"openstackclient\" (UID: \"383967e9-132f-4355-abbb-5904830137b7\") " pod="openstack/openstackclient" Mar 21 08:48:25 crc kubenswrapper[4696]: I0321 08:48:25.712500 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntzd6\" (UniqueName: \"kubernetes.io/projected/383967e9-132f-4355-abbb-5904830137b7-kube-api-access-ntzd6\") pod \"openstackclient\" (UID: \"383967e9-132f-4355-abbb-5904830137b7\") " pod="openstack/openstackclient" Mar 21 08:48:25 crc kubenswrapper[4696]: I0321 08:48:25.713369 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/383967e9-132f-4355-abbb-5904830137b7-openstack-config\") pod \"openstackclient\" (UID: \"383967e9-132f-4355-abbb-5904830137b7\") " pod="openstack/openstackclient" Mar 21 08:48:25 crc kubenswrapper[4696]: I0321 08:48:25.720106 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/383967e9-132f-4355-abbb-5904830137b7-combined-ca-bundle\") pod \"openstackclient\" (UID: \"383967e9-132f-4355-abbb-5904830137b7\") " pod="openstack/openstackclient" Mar 21 08:48:25 crc kubenswrapper[4696]: I0321 08:48:25.721135 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/383967e9-132f-4355-abbb-5904830137b7-openstack-config-secret\") pod \"openstackclient\" (UID: \"383967e9-132f-4355-abbb-5904830137b7\") " pod="openstack/openstackclient" Mar 21 08:48:25 crc kubenswrapper[4696]: I0321 08:48:25.727921 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntzd6\" (UniqueName: \"kubernetes.io/projected/383967e9-132f-4355-abbb-5904830137b7-kube-api-access-ntzd6\") pod \"openstackclient\" (UID: \"383967e9-132f-4355-abbb-5904830137b7\") " pod="openstack/openstackclient" Mar 21 08:48:25 crc kubenswrapper[4696]: I0321 08:48:25.839077 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 21 08:48:26 crc kubenswrapper[4696]: I0321 08:48:26.048513 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4fce4b85-2f39-4702-b221-fd5e870e6a2d","Type":"ContainerStarted","Data":"c27bbf6f73454c530e920b9a3d09f7f3e925f5effc025b966228d9003f91a64f"} Mar 21 08:48:26 crc kubenswrapper[4696]: I0321 08:48:26.488659 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.488637993 podStartE2EDuration="4.488637993s" podCreationTimestamp="2026-03-21 08:48:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:48:26.077464807 +0000 UTC m=+1240.198345520" watchObservedRunningTime="2026-03-21 08:48:26.488637993 +0000 UTC m=+1240.609518706" Mar 21 08:48:26 crc kubenswrapper[4696]: I0321 08:48:26.491719 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 21 08:48:26 crc kubenswrapper[4696]: W0321 08:48:26.499141 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod383967e9_132f_4355_abbb_5904830137b7.slice/crio-580459eb6d0e1f9b05c5129cb14c0674b3d22bb89200e340a1c34841ae1bae8c WatchSource:0}: Error finding container 580459eb6d0e1f9b05c5129cb14c0674b3d22bb89200e340a1c34841ae1bae8c: Status 404 returned error can't find the container with id 580459eb6d0e1f9b05c5129cb14c0674b3d22bb89200e340a1c34841ae1bae8c Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.060852 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"383967e9-132f-4355-abbb-5904830137b7","Type":"ContainerStarted","Data":"580459eb6d0e1f9b05c5129cb14c0674b3d22bb89200e340a1c34841ae1bae8c"} Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.346550 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7b65745bbc-5b5zz"] Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.348871 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.356274 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.356476 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.356578 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.392114 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7b65745bbc-5b5zz"] Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.451047 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-internal-tls-certs\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.451156 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-config-data\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.451192 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-combined-ca-bundle\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.451257 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-run-httpd\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.451450 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n24jc\" (UniqueName: \"kubernetes.io/projected/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-kube-api-access-n24jc\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.451748 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-log-httpd\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.451785 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-etc-swift\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.451850 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-public-tls-certs\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.553208 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-config-data\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.553549 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-combined-ca-bundle\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.553575 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-run-httpd\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.553626 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n24jc\" (UniqueName: \"kubernetes.io/projected/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-kube-api-access-n24jc\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.553709 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-log-httpd\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.553728 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-etc-swift\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.553748 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-public-tls-certs\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.553770 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-internal-tls-certs\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.554163 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-run-httpd\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.554236 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-log-httpd\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.561893 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-combined-ca-bundle\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.562375 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-internal-tls-certs\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.562566 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-config-data\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.563228 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-etc-swift\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.569258 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-public-tls-certs\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.578522 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n24jc\" (UniqueName: \"kubernetes.io/projected/4abc84ca-2aab-4423-b934-c61f3ef8ea6d-kube-api-access-n24jc\") pod \"swift-proxy-7b65745bbc-5b5zz\" (UID: \"4abc84ca-2aab-4423-b934-c61f3ef8ea6d\") " pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.706146 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.775779 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.824159 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-56b7788dc4-f928g" Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.915566 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-565c47bc8-bmdxv"] Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.915842 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-565c47bc8-bmdxv" podUID="a32ac807-7955-425a-a971-e07ddde36bb0" containerName="placement-log" containerID="cri-o://55862ad01a19e70bcd5fe3ea21ef73df613a89f8d76d7c0d42d3c99ae1f5d4db" gracePeriod=30 Mar 21 08:48:27 crc kubenswrapper[4696]: I0321 08:48:27.916281 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-565c47bc8-bmdxv" podUID="a32ac807-7955-425a-a971-e07ddde36bb0" containerName="placement-api" containerID="cri-o://753467b400aac31c890d0990cf0105c1387d9e9a97ee9e201734937908aa8474" gracePeriod=30 Mar 21 08:48:28 crc kubenswrapper[4696]: I0321 08:48:28.080333 4696 generic.go:334] "Generic (PLEG): container finished" podID="a32ac807-7955-425a-a971-e07ddde36bb0" containerID="55862ad01a19e70bcd5fe3ea21ef73df613a89f8d76d7c0d42d3c99ae1f5d4db" exitCode=143 Mar 21 08:48:28 crc kubenswrapper[4696]: I0321 08:48:28.080871 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-565c47bc8-bmdxv" event={"ID":"a32ac807-7955-425a-a971-e07ddde36bb0","Type":"ContainerDied","Data":"55862ad01a19e70bcd5fe3ea21ef73df613a89f8d76d7c0d42d3c99ae1f5d4db"} Mar 21 08:48:28 crc kubenswrapper[4696]: I0321 08:48:28.400144 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 21 08:48:28 crc kubenswrapper[4696]: I0321 08:48:28.494332 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7b65745bbc-5b5zz"] Mar 21 08:48:28 crc kubenswrapper[4696]: W0321 08:48:28.512973 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4abc84ca_2aab_4423_b934_c61f3ef8ea6d.slice/crio-315cb8ae905c9d9e6da49f74d9eb35cdae6ae41cde96d22811682cea6ab9c85d WatchSource:0}: Error finding container 315cb8ae905c9d9e6da49f74d9eb35cdae6ae41cde96d22811682cea6ab9c85d: Status 404 returned error can't find the container with id 315cb8ae905c9d9e6da49f74d9eb35cdae6ae41cde96d22811682cea6ab9c85d Mar 21 08:48:29 crc kubenswrapper[4696]: I0321 08:48:29.095493 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b65745bbc-5b5zz" event={"ID":"4abc84ca-2aab-4423-b934-c61f3ef8ea6d","Type":"ContainerStarted","Data":"43f389db563c661bdc95528365be74a1b4c8a4dc3ede8ae28a650904daab15a3"} Mar 21 08:48:29 crc kubenswrapper[4696]: I0321 08:48:29.095842 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b65745bbc-5b5zz" event={"ID":"4abc84ca-2aab-4423-b934-c61f3ef8ea6d","Type":"ContainerStarted","Data":"315cb8ae905c9d9e6da49f74d9eb35cdae6ae41cde96d22811682cea6ab9c85d"} Mar 21 08:48:29 crc kubenswrapper[4696]: I0321 08:48:29.136758 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 21 08:48:29 crc kubenswrapper[4696]: I0321 08:48:29.413648 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:48:29 crc kubenswrapper[4696]: I0321 08:48:29.414135 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1e007714-252e-43e9-b281-323db9d580ad" containerName="ceilometer-central-agent" containerID="cri-o://fa78034c74c9f61a2e6cff2f20c9d14454749e1e7b00d112e2dadbe48a3560cb" gracePeriod=30 Mar 21 08:48:29 crc kubenswrapper[4696]: I0321 08:48:29.414186 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1e007714-252e-43e9-b281-323db9d580ad" containerName="ceilometer-notification-agent" containerID="cri-o://1ec234430915a7754af8a24e7b6a89c1a27fbe794f01a4f8c0b2b4cc619c41bf" gracePeriod=30 Mar 21 08:48:29 crc kubenswrapper[4696]: I0321 08:48:29.414185 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1e007714-252e-43e9-b281-323db9d580ad" containerName="proxy-httpd" containerID="cri-o://672e68b1de41b8ea639fbd69f58f1e6062fbf25d20341f77c67fd93a4b9c35a7" gracePeriod=30 Mar 21 08:48:29 crc kubenswrapper[4696]: I0321 08:48:29.414193 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1e007714-252e-43e9-b281-323db9d580ad" containerName="sg-core" containerID="cri-o://0f683af8454805602847c08cf6ff20f6a5a145e396e630d81e262794d191a866" gracePeriod=30 Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.130527 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b65745bbc-5b5zz" event={"ID":"4abc84ca-2aab-4423-b934-c61f3ef8ea6d","Type":"ContainerStarted","Data":"6c30f389267f40a872626825bc7f4537339ac8e192c07d77010a4c74721c3803"} Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.130867 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.130890 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.151895 4696 generic.go:334] "Generic (PLEG): container finished" podID="1e007714-252e-43e9-b281-323db9d580ad" containerID="672e68b1de41b8ea639fbd69f58f1e6062fbf25d20341f77c67fd93a4b9c35a7" exitCode=0 Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.151931 4696 generic.go:334] "Generic (PLEG): container finished" podID="1e007714-252e-43e9-b281-323db9d580ad" containerID="0f683af8454805602847c08cf6ff20f6a5a145e396e630d81e262794d191a866" exitCode=2 Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.151939 4696 generic.go:334] "Generic (PLEG): container finished" podID="1e007714-252e-43e9-b281-323db9d580ad" containerID="1ec234430915a7754af8a24e7b6a89c1a27fbe794f01a4f8c0b2b4cc619c41bf" exitCode=0 Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.151946 4696 generic.go:334] "Generic (PLEG): container finished" podID="1e007714-252e-43e9-b281-323db9d580ad" containerID="fa78034c74c9f61a2e6cff2f20c9d14454749e1e7b00d112e2dadbe48a3560cb" exitCode=0 Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.151968 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e007714-252e-43e9-b281-323db9d580ad","Type":"ContainerDied","Data":"672e68b1de41b8ea639fbd69f58f1e6062fbf25d20341f77c67fd93a4b9c35a7"} Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.151993 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e007714-252e-43e9-b281-323db9d580ad","Type":"ContainerDied","Data":"0f683af8454805602847c08cf6ff20f6a5a145e396e630d81e262794d191a866"} Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.152002 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e007714-252e-43e9-b281-323db9d580ad","Type":"ContainerDied","Data":"1ec234430915a7754af8a24e7b6a89c1a27fbe794f01a4f8c0b2b4cc619c41bf"} Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.152011 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e007714-252e-43e9-b281-323db9d580ad","Type":"ContainerDied","Data":"fa78034c74c9f61a2e6cff2f20c9d14454749e1e7b00d112e2dadbe48a3560cb"} Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.152293 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7b65745bbc-5b5zz" podStartSLOduration=3.152274055 podStartE2EDuration="3.152274055s" podCreationTimestamp="2026-03-21 08:48:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:48:30.151450693 +0000 UTC m=+1244.272331416" watchObservedRunningTime="2026-03-21 08:48:30.152274055 +0000 UTC m=+1244.273154768" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.262657 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.347609 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-8w2b4"] Mar 21 08:48:30 crc kubenswrapper[4696]: E0321 08:48:30.348057 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e007714-252e-43e9-b281-323db9d580ad" containerName="proxy-httpd" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.348075 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e007714-252e-43e9-b281-323db9d580ad" containerName="proxy-httpd" Mar 21 08:48:30 crc kubenswrapper[4696]: E0321 08:48:30.348087 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e007714-252e-43e9-b281-323db9d580ad" containerName="ceilometer-notification-agent" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.348094 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e007714-252e-43e9-b281-323db9d580ad" containerName="ceilometer-notification-agent" Mar 21 08:48:30 crc kubenswrapper[4696]: E0321 08:48:30.348107 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e007714-252e-43e9-b281-323db9d580ad" containerName="ceilometer-central-agent" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.348113 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e007714-252e-43e9-b281-323db9d580ad" containerName="ceilometer-central-agent" Mar 21 08:48:30 crc kubenswrapper[4696]: E0321 08:48:30.348141 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e007714-252e-43e9-b281-323db9d580ad" containerName="sg-core" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.348147 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e007714-252e-43e9-b281-323db9d580ad" containerName="sg-core" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.349428 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e007714-252e-43e9-b281-323db9d580ad" containerName="sg-core" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.349459 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e007714-252e-43e9-b281-323db9d580ad" containerName="ceilometer-central-agent" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.349475 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e007714-252e-43e9-b281-323db9d580ad" containerName="proxy-httpd" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.349499 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e007714-252e-43e9-b281-323db9d580ad" containerName="ceilometer-notification-agent" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.350303 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8w2b4" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.363133 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8w2b4"] Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.438240 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e007714-252e-43e9-b281-323db9d580ad-log-httpd\") pod \"1e007714-252e-43e9-b281-323db9d580ad\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.438320 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-sg-core-conf-yaml\") pod \"1e007714-252e-43e9-b281-323db9d580ad\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.438407 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-config-data\") pod \"1e007714-252e-43e9-b281-323db9d580ad\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.438425 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-combined-ca-bundle\") pod \"1e007714-252e-43e9-b281-323db9d580ad\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.438528 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e007714-252e-43e9-b281-323db9d580ad-run-httpd\") pod \"1e007714-252e-43e9-b281-323db9d580ad\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.438566 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kznqc\" (UniqueName: \"kubernetes.io/projected/1e007714-252e-43e9-b281-323db9d580ad-kube-api-access-kznqc\") pod \"1e007714-252e-43e9-b281-323db9d580ad\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.438592 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-scripts\") pod \"1e007714-252e-43e9-b281-323db9d580ad\" (UID: \"1e007714-252e-43e9-b281-323db9d580ad\") " Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.438701 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e007714-252e-43e9-b281-323db9d580ad-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1e007714-252e-43e9-b281-323db9d580ad" (UID: "1e007714-252e-43e9-b281-323db9d580ad"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.438849 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e007714-252e-43e9-b281-323db9d580ad-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1e007714-252e-43e9-b281-323db9d580ad" (UID: "1e007714-252e-43e9-b281-323db9d580ad"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.439421 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cpss\" (UniqueName: \"kubernetes.io/projected/2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d-kube-api-access-4cpss\") pod \"nova-api-db-create-8w2b4\" (UID: \"2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d\") " pod="openstack/nova-api-db-create-8w2b4" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.439450 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d-operator-scripts\") pod \"nova-api-db-create-8w2b4\" (UID: \"2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d\") " pod="openstack/nova-api-db-create-8w2b4" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.439525 4696 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e007714-252e-43e9-b281-323db9d580ad-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.439537 4696 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e007714-252e-43e9-b281-323db9d580ad-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.461562 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-scripts" (OuterVolumeSpecName: "scripts") pod "1e007714-252e-43e9-b281-323db9d580ad" (UID: "1e007714-252e-43e9-b281-323db9d580ad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.461690 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e007714-252e-43e9-b281-323db9d580ad-kube-api-access-kznqc" (OuterVolumeSpecName: "kube-api-access-kznqc") pod "1e007714-252e-43e9-b281-323db9d580ad" (UID: "1e007714-252e-43e9-b281-323db9d580ad"). InnerVolumeSpecName "kube-api-access-kznqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.469770 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1e007714-252e-43e9-b281-323db9d580ad" (UID: "1e007714-252e-43e9-b281-323db9d580ad"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.543243 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cpss\" (UniqueName: \"kubernetes.io/projected/2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d-kube-api-access-4cpss\") pod \"nova-api-db-create-8w2b4\" (UID: \"2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d\") " pod="openstack/nova-api-db-create-8w2b4" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.543283 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d-operator-scripts\") pod \"nova-api-db-create-8w2b4\" (UID: \"2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d\") " pod="openstack/nova-api-db-create-8w2b4" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.543589 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kznqc\" (UniqueName: \"kubernetes.io/projected/1e007714-252e-43e9-b281-323db9d580ad-kube-api-access-kznqc\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.543604 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.543614 4696 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.544236 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d-operator-scripts\") pod \"nova-api-db-create-8w2b4\" (UID: \"2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d\") " pod="openstack/nova-api-db-create-8w2b4" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.578244 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cpss\" (UniqueName: \"kubernetes.io/projected/2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d-kube-api-access-4cpss\") pod \"nova-api-db-create-8w2b4\" (UID: \"2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d\") " pod="openstack/nova-api-db-create-8w2b4" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.648123 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e007714-252e-43e9-b281-323db9d580ad" (UID: "1e007714-252e-43e9-b281-323db9d580ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.649729 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.663273 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-config-data" (OuterVolumeSpecName: "config-data") pod "1e007714-252e-43e9-b281-323db9d580ad" (UID: "1e007714-252e-43e9-b281-323db9d580ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.670597 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8w2b4" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.752998 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e007714-252e-43e9-b281-323db9d580ad-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.816789 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-p6m2c"] Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.818077 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-297e-account-create-update-9rqlh"] Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.818979 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-p6m2c"] Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.818995 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-297e-account-create-update-9rqlh"] Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.819006 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-4t9zv"] Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.819954 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-4t9zv"] Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.819976 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-9b07-account-create-update-g6ddw"] Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.820856 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9b07-account-create-update-g6ddw" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.820979 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-297e-account-create-update-9rqlh" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.821033 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4t9zv" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.820993 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-p6m2c" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.826154 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.826384 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.839964 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9b07-account-create-update-g6ddw"] Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.978869 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b58tp\" (UniqueName: \"kubernetes.io/projected/54341261-11d8-407a-a339-6c0d70c1ad5f-kube-api-access-b58tp\") pod \"nova-cell0-9b07-account-create-update-g6ddw\" (UID: \"54341261-11d8-407a-a339-6c0d70c1ad5f\") " pod="openstack/nova-cell0-9b07-account-create-update-g6ddw" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.979200 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/092019b5-fac9-448e-be38-7b1cbc62f107-operator-scripts\") pod \"nova-cell0-db-create-p6m2c\" (UID: \"092019b5-fac9-448e-be38-7b1cbc62f107\") " pod="openstack/nova-cell0-db-create-p6m2c" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.979313 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrqs8\" (UniqueName: \"kubernetes.io/projected/092019b5-fac9-448e-be38-7b1cbc62f107-kube-api-access-nrqs8\") pod \"nova-cell0-db-create-p6m2c\" (UID: \"092019b5-fac9-448e-be38-7b1cbc62f107\") " pod="openstack/nova-cell0-db-create-p6m2c" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.979387 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgqqt\" (UniqueName: \"kubernetes.io/projected/582c3a4a-774e-4748-b055-6e2dcf1065af-kube-api-access-vgqqt\") pod \"nova-cell1-db-create-4t9zv\" (UID: \"582c3a4a-774e-4748-b055-6e2dcf1065af\") " pod="openstack/nova-cell1-db-create-4t9zv" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.979422 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/582c3a4a-774e-4748-b055-6e2dcf1065af-operator-scripts\") pod \"nova-cell1-db-create-4t9zv\" (UID: \"582c3a4a-774e-4748-b055-6e2dcf1065af\") " pod="openstack/nova-cell1-db-create-4t9zv" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.979607 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2-operator-scripts\") pod \"nova-api-297e-account-create-update-9rqlh\" (UID: \"61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2\") " pod="openstack/nova-api-297e-account-create-update-9rqlh" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.979694 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54341261-11d8-407a-a339-6c0d70c1ad5f-operator-scripts\") pod \"nova-cell0-9b07-account-create-update-g6ddw\" (UID: \"54341261-11d8-407a-a339-6c0d70c1ad5f\") " pod="openstack/nova-cell0-9b07-account-create-update-g6ddw" Mar 21 08:48:30 crc kubenswrapper[4696]: I0321 08:48:30.979908 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxjg5\" (UniqueName: \"kubernetes.io/projected/61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2-kube-api-access-wxjg5\") pod \"nova-api-297e-account-create-update-9rqlh\" (UID: \"61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2\") " pod="openstack/nova-api-297e-account-create-update-9rqlh" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.052738 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-9172-account-create-update-r4fhw"] Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.063239 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9172-account-create-update-r4fhw" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.065718 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.088449 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-9172-account-create-update-r4fhw"] Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.098937 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54341261-11d8-407a-a339-6c0d70c1ad5f-operator-scripts\") pod \"nova-cell0-9b07-account-create-update-g6ddw\" (UID: \"54341261-11d8-407a-a339-6c0d70c1ad5f\") " pod="openstack/nova-cell0-9b07-account-create-update-g6ddw" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.099032 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxjg5\" (UniqueName: \"kubernetes.io/projected/61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2-kube-api-access-wxjg5\") pod \"nova-api-297e-account-create-update-9rqlh\" (UID: \"61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2\") " pod="openstack/nova-api-297e-account-create-update-9rqlh" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.099078 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b58tp\" (UniqueName: \"kubernetes.io/projected/54341261-11d8-407a-a339-6c0d70c1ad5f-kube-api-access-b58tp\") pod \"nova-cell0-9b07-account-create-update-g6ddw\" (UID: \"54341261-11d8-407a-a339-6c0d70c1ad5f\") " pod="openstack/nova-cell0-9b07-account-create-update-g6ddw" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.099177 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/092019b5-fac9-448e-be38-7b1cbc62f107-operator-scripts\") pod \"nova-cell0-db-create-p6m2c\" (UID: \"092019b5-fac9-448e-be38-7b1cbc62f107\") " pod="openstack/nova-cell0-db-create-p6m2c" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.099196 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrqs8\" (UniqueName: \"kubernetes.io/projected/092019b5-fac9-448e-be38-7b1cbc62f107-kube-api-access-nrqs8\") pod \"nova-cell0-db-create-p6m2c\" (UID: \"092019b5-fac9-448e-be38-7b1cbc62f107\") " pod="openstack/nova-cell0-db-create-p6m2c" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.099233 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgqqt\" (UniqueName: \"kubernetes.io/projected/582c3a4a-774e-4748-b055-6e2dcf1065af-kube-api-access-vgqqt\") pod \"nova-cell1-db-create-4t9zv\" (UID: \"582c3a4a-774e-4748-b055-6e2dcf1065af\") " pod="openstack/nova-cell1-db-create-4t9zv" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.099311 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/582c3a4a-774e-4748-b055-6e2dcf1065af-operator-scripts\") pod \"nova-cell1-db-create-4t9zv\" (UID: \"582c3a4a-774e-4748-b055-6e2dcf1065af\") " pod="openstack/nova-cell1-db-create-4t9zv" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.099356 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2-operator-scripts\") pod \"nova-api-297e-account-create-update-9rqlh\" (UID: \"61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2\") " pod="openstack/nova-api-297e-account-create-update-9rqlh" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.100176 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2-operator-scripts\") pod \"nova-api-297e-account-create-update-9rqlh\" (UID: \"61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2\") " pod="openstack/nova-api-297e-account-create-update-9rqlh" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.100665 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54341261-11d8-407a-a339-6c0d70c1ad5f-operator-scripts\") pod \"nova-cell0-9b07-account-create-update-g6ddw\" (UID: \"54341261-11d8-407a-a339-6c0d70c1ad5f\") " pod="openstack/nova-cell0-9b07-account-create-update-g6ddw" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.103762 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/582c3a4a-774e-4748-b055-6e2dcf1065af-operator-scripts\") pod \"nova-cell1-db-create-4t9zv\" (UID: \"582c3a4a-774e-4748-b055-6e2dcf1065af\") " pod="openstack/nova-cell1-db-create-4t9zv" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.129790 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrqs8\" (UniqueName: \"kubernetes.io/projected/092019b5-fac9-448e-be38-7b1cbc62f107-kube-api-access-nrqs8\") pod \"nova-cell0-db-create-p6m2c\" (UID: \"092019b5-fac9-448e-be38-7b1cbc62f107\") " pod="openstack/nova-cell0-db-create-p6m2c" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.137442 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/092019b5-fac9-448e-be38-7b1cbc62f107-operator-scripts\") pod \"nova-cell0-db-create-p6m2c\" (UID: \"092019b5-fac9-448e-be38-7b1cbc62f107\") " pod="openstack/nova-cell0-db-create-p6m2c" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.139357 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b58tp\" (UniqueName: \"kubernetes.io/projected/54341261-11d8-407a-a339-6c0d70c1ad5f-kube-api-access-b58tp\") pod \"nova-cell0-9b07-account-create-update-g6ddw\" (UID: \"54341261-11d8-407a-a339-6c0d70c1ad5f\") " pod="openstack/nova-cell0-9b07-account-create-update-g6ddw" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.139439 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgqqt\" (UniqueName: \"kubernetes.io/projected/582c3a4a-774e-4748-b055-6e2dcf1065af-kube-api-access-vgqqt\") pod \"nova-cell1-db-create-4t9zv\" (UID: \"582c3a4a-774e-4748-b055-6e2dcf1065af\") " pod="openstack/nova-cell1-db-create-4t9zv" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.153529 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxjg5\" (UniqueName: \"kubernetes.io/projected/61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2-kube-api-access-wxjg5\") pod \"nova-api-297e-account-create-update-9rqlh\" (UID: \"61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2\") " pod="openstack/nova-api-297e-account-create-update-9rqlh" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.175567 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9b07-account-create-update-g6ddw" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.201380 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9fe5e80-cc65-4225-9b74-73bf39a0f187-operator-scripts\") pod \"nova-cell1-9172-account-create-update-r4fhw\" (UID: \"b9fe5e80-cc65-4225-9b74-73bf39a0f187\") " pod="openstack/nova-cell1-9172-account-create-update-r4fhw" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.201456 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c45wf\" (UniqueName: \"kubernetes.io/projected/b9fe5e80-cc65-4225-9b74-73bf39a0f187-kube-api-access-c45wf\") pod \"nova-cell1-9172-account-create-update-r4fhw\" (UID: \"b9fe5e80-cc65-4225-9b74-73bf39a0f187\") " pod="openstack/nova-cell1-9172-account-create-update-r4fhw" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.203385 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-297e-account-create-update-9rqlh" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.228280 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4t9zv" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.244294 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-p6m2c" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.255889 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.255945 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e007714-252e-43e9-b281-323db9d580ad","Type":"ContainerDied","Data":"27d84e25c5928559268ad7a13b8eba6d0a7c02093f825ef16ae5fa6da1d8e909"} Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.256005 4696 scope.go:117] "RemoveContainer" containerID="672e68b1de41b8ea639fbd69f58f1e6062fbf25d20341f77c67fd93a4b9c35a7" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.316180 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9fe5e80-cc65-4225-9b74-73bf39a0f187-operator-scripts\") pod \"nova-cell1-9172-account-create-update-r4fhw\" (UID: \"b9fe5e80-cc65-4225-9b74-73bf39a0f187\") " pod="openstack/nova-cell1-9172-account-create-update-r4fhw" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.316240 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c45wf\" (UniqueName: \"kubernetes.io/projected/b9fe5e80-cc65-4225-9b74-73bf39a0f187-kube-api-access-c45wf\") pod \"nova-cell1-9172-account-create-update-r4fhw\" (UID: \"b9fe5e80-cc65-4225-9b74-73bf39a0f187\") " pod="openstack/nova-cell1-9172-account-create-update-r4fhw" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.317691 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9fe5e80-cc65-4225-9b74-73bf39a0f187-operator-scripts\") pod \"nova-cell1-9172-account-create-update-r4fhw\" (UID: \"b9fe5e80-cc65-4225-9b74-73bf39a0f187\") " pod="openstack/nova-cell1-9172-account-create-update-r4fhw" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.399530 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c45wf\" (UniqueName: \"kubernetes.io/projected/b9fe5e80-cc65-4225-9b74-73bf39a0f187-kube-api-access-c45wf\") pod \"nova-cell1-9172-account-create-update-r4fhw\" (UID: \"b9fe5e80-cc65-4225-9b74-73bf39a0f187\") " pod="openstack/nova-cell1-9172-account-create-update-r4fhw" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.418964 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.430558 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.446737 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.449395 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.456340 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.456542 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.460071 4696 scope.go:117] "RemoveContainer" containerID="0f683af8454805602847c08cf6ff20f6a5a145e396e630d81e262794d191a866" Mar 21 08:48:31 crc kubenswrapper[4696]: W0321 08:48:31.477324 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2aefd2d5_17ac_43ef_b7ac_a06f7b46ed6d.slice/crio-e937813d81562d807962c413afb6ad61a6e55b676a7efcfae00000415e94507f WatchSource:0}: Error finding container e937813d81562d807962c413afb6ad61a6e55b676a7efcfae00000415e94507f: Status 404 returned error can't find the container with id e937813d81562d807962c413afb6ad61a6e55b676a7efcfae00000415e94507f Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.478312 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8w2b4"] Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.504471 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.511443 4696 scope.go:117] "RemoveContainer" containerID="1ec234430915a7754af8a24e7b6a89c1a27fbe794f01a4f8c0b2b4cc619c41bf" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.626106 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xml8b\" (UniqueName: \"kubernetes.io/projected/1195f5f9-d86b-4038-8ce9-1573e13ed98e-kube-api-access-xml8b\") pod \"ceilometer-0\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.626301 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1195f5f9-d86b-4038-8ce9-1573e13ed98e-log-httpd\") pod \"ceilometer-0\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.626448 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-config-data\") pod \"ceilometer-0\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.626498 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.626554 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1195f5f9-d86b-4038-8ce9-1573e13ed98e-run-httpd\") pod \"ceilometer-0\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.626703 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.626815 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-scripts\") pod \"ceilometer-0\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.698672 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9172-account-create-update-r4fhw" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.728442 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1195f5f9-d86b-4038-8ce9-1573e13ed98e-run-httpd\") pod \"ceilometer-0\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.728580 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.728623 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-scripts\") pod \"ceilometer-0\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.728696 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xml8b\" (UniqueName: \"kubernetes.io/projected/1195f5f9-d86b-4038-8ce9-1573e13ed98e-kube-api-access-xml8b\") pod \"ceilometer-0\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.728767 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1195f5f9-d86b-4038-8ce9-1573e13ed98e-log-httpd\") pod \"ceilometer-0\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.728868 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-config-data\") pod \"ceilometer-0\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.728924 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.730447 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1195f5f9-d86b-4038-8ce9-1573e13ed98e-log-httpd\") pod \"ceilometer-0\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.730962 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1195f5f9-d86b-4038-8ce9-1573e13ed98e-run-httpd\") pod \"ceilometer-0\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.737215 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-config-data\") pod \"ceilometer-0\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.742705 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-scripts\") pod \"ceilometer-0\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.748557 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.752554 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xml8b\" (UniqueName: \"kubernetes.io/projected/1195f5f9-d86b-4038-8ce9-1573e13ed98e-kube-api-access-xml8b\") pod \"ceilometer-0\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.753374 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " pod="openstack/ceilometer-0" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.771997 4696 scope.go:117] "RemoveContainer" containerID="fa78034c74c9f61a2e6cff2f20c9d14454749e1e7b00d112e2dadbe48a3560cb" Mar 21 08:48:31 crc kubenswrapper[4696]: I0321 08:48:31.790208 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.105751 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.199641 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-4t9zv"] Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.213729 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-297e-account-create-update-9rqlh"] Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.240202 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-combined-ca-bundle\") pod \"a32ac807-7955-425a-a971-e07ddde36bb0\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.240271 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-public-tls-certs\") pod \"a32ac807-7955-425a-a971-e07ddde36bb0\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.240407 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-internal-tls-certs\") pod \"a32ac807-7955-425a-a971-e07ddde36bb0\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.240524 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-config-data\") pod \"a32ac807-7955-425a-a971-e07ddde36bb0\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.240624 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a32ac807-7955-425a-a971-e07ddde36bb0-logs\") pod \"a32ac807-7955-425a-a971-e07ddde36bb0\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.240670 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xk4pp\" (UniqueName: \"kubernetes.io/projected/a32ac807-7955-425a-a971-e07ddde36bb0-kube-api-access-xk4pp\") pod \"a32ac807-7955-425a-a971-e07ddde36bb0\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.240690 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-scripts\") pod \"a32ac807-7955-425a-a971-e07ddde36bb0\" (UID: \"a32ac807-7955-425a-a971-e07ddde36bb0\") " Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.243115 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a32ac807-7955-425a-a971-e07ddde36bb0-logs" (OuterVolumeSpecName: "logs") pod "a32ac807-7955-425a-a971-e07ddde36bb0" (UID: "a32ac807-7955-425a-a971-e07ddde36bb0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.245577 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-scripts" (OuterVolumeSpecName: "scripts") pod "a32ac807-7955-425a-a971-e07ddde36bb0" (UID: "a32ac807-7955-425a-a971-e07ddde36bb0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.252566 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a32ac807-7955-425a-a971-e07ddde36bb0-kube-api-access-xk4pp" (OuterVolumeSpecName: "kube-api-access-xk4pp") pod "a32ac807-7955-425a-a971-e07ddde36bb0" (UID: "a32ac807-7955-425a-a971-e07ddde36bb0"). InnerVolumeSpecName "kube-api-access-xk4pp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.290724 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8w2b4" event={"ID":"2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d","Type":"ContainerStarted","Data":"d691cb047a6743e59099baef4ff05d5f1de76cddf6715b786b4e07e0cc1eb607"} Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.290771 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8w2b4" event={"ID":"2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d","Type":"ContainerStarted","Data":"e937813d81562d807962c413afb6ad61a6e55b676a7efcfae00000415e94507f"} Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.296252 4696 generic.go:334] "Generic (PLEG): container finished" podID="a32ac807-7955-425a-a971-e07ddde36bb0" containerID="753467b400aac31c890d0990cf0105c1387d9e9a97ee9e201734937908aa8474" exitCode=0 Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.296329 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-565c47bc8-bmdxv" event={"ID":"a32ac807-7955-425a-a971-e07ddde36bb0","Type":"ContainerDied","Data":"753467b400aac31c890d0990cf0105c1387d9e9a97ee9e201734937908aa8474"} Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.296364 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-565c47bc8-bmdxv" event={"ID":"a32ac807-7955-425a-a971-e07ddde36bb0","Type":"ContainerDied","Data":"5c2afb4660e62cf6acd556874a3e10004cdcf9eb8e4a65657b808244452e0d04"} Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.296385 4696 scope.go:117] "RemoveContainer" containerID="753467b400aac31c890d0990cf0105c1387d9e9a97ee9e201734937908aa8474" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.296541 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-565c47bc8-bmdxv" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.337626 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4t9zv" event={"ID":"582c3a4a-774e-4748-b055-6e2dcf1065af","Type":"ContainerStarted","Data":"0d20eac6aa5e4b57ef71b33daf3a22a7704daf6292b38b8abda0274d8ba92e92"} Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.343026 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a32ac807-7955-425a-a971-e07ddde36bb0-logs\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.345629 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xk4pp\" (UniqueName: \"kubernetes.io/projected/a32ac807-7955-425a-a971-e07ddde36bb0-kube-api-access-xk4pp\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.345724 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.364320 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-8w2b4" podStartSLOduration=2.36430318 podStartE2EDuration="2.36430318s" podCreationTimestamp="2026-03-21 08:48:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:48:32.338182477 +0000 UTC m=+1246.459063190" watchObservedRunningTime="2026-03-21 08:48:32.36430318 +0000 UTC m=+1246.485183893" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.366073 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-297e-account-create-update-9rqlh" event={"ID":"61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2","Type":"ContainerStarted","Data":"72108c518b505ef09c4229ffcd256bc4ffa32b056cebbf42ee01277e7d906211"} Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.381677 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-config-data" (OuterVolumeSpecName: "config-data") pod "a32ac807-7955-425a-a971-e07ddde36bb0" (UID: "a32ac807-7955-425a-a971-e07ddde36bb0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.391767 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-p6m2c"] Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.398525 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a32ac807-7955-425a-a971-e07ddde36bb0" (UID: "a32ac807-7955-425a-a971-e07ddde36bb0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.407342 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9b07-account-create-update-g6ddw"] Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.445506 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a32ac807-7955-425a-a971-e07ddde36bb0" (UID: "a32ac807-7955-425a-a971-e07ddde36bb0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.449448 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.449473 4696 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.449483 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.458283 4696 scope.go:117] "RemoveContainer" containerID="55862ad01a19e70bcd5fe3ea21ef73df613a89f8d76d7c0d42d3c99ae1f5d4db" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.547555 4696 scope.go:117] "RemoveContainer" containerID="753467b400aac31c890d0990cf0105c1387d9e9a97ee9e201734937908aa8474" Mar 21 08:48:32 crc kubenswrapper[4696]: E0321 08:48:32.551779 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"753467b400aac31c890d0990cf0105c1387d9e9a97ee9e201734937908aa8474\": container with ID starting with 753467b400aac31c890d0990cf0105c1387d9e9a97ee9e201734937908aa8474 not found: ID does not exist" containerID="753467b400aac31c890d0990cf0105c1387d9e9a97ee9e201734937908aa8474" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.551835 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"753467b400aac31c890d0990cf0105c1387d9e9a97ee9e201734937908aa8474"} err="failed to get container status \"753467b400aac31c890d0990cf0105c1387d9e9a97ee9e201734937908aa8474\": rpc error: code = NotFound desc = could not find container \"753467b400aac31c890d0990cf0105c1387d9e9a97ee9e201734937908aa8474\": container with ID starting with 753467b400aac31c890d0990cf0105c1387d9e9a97ee9e201734937908aa8474 not found: ID does not exist" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.551864 4696 scope.go:117] "RemoveContainer" containerID="55862ad01a19e70bcd5fe3ea21ef73df613a89f8d76d7c0d42d3c99ae1f5d4db" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.551949 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a32ac807-7955-425a-a971-e07ddde36bb0" (UID: "a32ac807-7955-425a-a971-e07ddde36bb0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.553310 4696 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a32ac807-7955-425a-a971-e07ddde36bb0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:32 crc kubenswrapper[4696]: E0321 08:48:32.560849 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55862ad01a19e70bcd5fe3ea21ef73df613a89f8d76d7c0d42d3c99ae1f5d4db\": container with ID starting with 55862ad01a19e70bcd5fe3ea21ef73df613a89f8d76d7c0d42d3c99ae1f5d4db not found: ID does not exist" containerID="55862ad01a19e70bcd5fe3ea21ef73df613a89f8d76d7c0d42d3c99ae1f5d4db" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.560889 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55862ad01a19e70bcd5fe3ea21ef73df613a89f8d76d7c0d42d3c99ae1f5d4db"} err="failed to get container status \"55862ad01a19e70bcd5fe3ea21ef73df613a89f8d76d7c0d42d3c99ae1f5d4db\": rpc error: code = NotFound desc = could not find container \"55862ad01a19e70bcd5fe3ea21ef73df613a89f8d76d7c0d42d3c99ae1f5d4db\": container with ID starting with 55862ad01a19e70bcd5fe3ea21ef73df613a89f8d76d7c0d42d3c99ae1f5d4db not found: ID does not exist" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.587211 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e007714-252e-43e9-b281-323db9d580ad" path="/var/lib/kubelet/pods/1e007714-252e-43e9-b281-323db9d580ad/volumes" Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.588229 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-9172-account-create-update-r4fhw"] Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.588280 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.757386 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-565c47bc8-bmdxv"] Mar 21 08:48:32 crc kubenswrapper[4696]: I0321 08:48:32.774339 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-565c47bc8-bmdxv"] Mar 21 08:48:33 crc kubenswrapper[4696]: I0321 08:48:33.376846 4696 generic.go:334] "Generic (PLEG): container finished" podID="2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d" containerID="d691cb047a6743e59099baef4ff05d5f1de76cddf6715b786b4e07e0cc1eb607" exitCode=0 Mar 21 08:48:33 crc kubenswrapper[4696]: I0321 08:48:33.376919 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8w2b4" event={"ID":"2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d","Type":"ContainerDied","Data":"d691cb047a6743e59099baef4ff05d5f1de76cddf6715b786b4e07e0cc1eb607"} Mar 21 08:48:33 crc kubenswrapper[4696]: I0321 08:48:33.379295 4696 generic.go:334] "Generic (PLEG): container finished" podID="b9fe5e80-cc65-4225-9b74-73bf39a0f187" containerID="115b17be47f80234d621a6e9d263a7d4f73af5b708364bcc07917182ddced170" exitCode=0 Mar 21 08:48:33 crc kubenswrapper[4696]: I0321 08:48:33.379355 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9172-account-create-update-r4fhw" event={"ID":"b9fe5e80-cc65-4225-9b74-73bf39a0f187","Type":"ContainerDied","Data":"115b17be47f80234d621a6e9d263a7d4f73af5b708364bcc07917182ddced170"} Mar 21 08:48:33 crc kubenswrapper[4696]: I0321 08:48:33.379457 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9172-account-create-update-r4fhw" event={"ID":"b9fe5e80-cc65-4225-9b74-73bf39a0f187","Type":"ContainerStarted","Data":"d7d6c6438f4c1c2db6139096858304266a0d12ef9a8863f4abc0714c6ff65493"} Mar 21 08:48:33 crc kubenswrapper[4696]: I0321 08:48:33.382370 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1195f5f9-d86b-4038-8ce9-1573e13ed98e","Type":"ContainerStarted","Data":"7fe3c8481a51d7859bef649640eb629a0eb5cae6a143aacf8f505e978df280ed"} Mar 21 08:48:33 crc kubenswrapper[4696]: I0321 08:48:33.383607 4696 generic.go:334] "Generic (PLEG): container finished" podID="582c3a4a-774e-4748-b055-6e2dcf1065af" containerID="a01d9d82b3cfea30bfced4f72a1e28c6089568366cfb2d31ad682c7a39cec486" exitCode=0 Mar 21 08:48:33 crc kubenswrapper[4696]: I0321 08:48:33.383666 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4t9zv" event={"ID":"582c3a4a-774e-4748-b055-6e2dcf1065af","Type":"ContainerDied","Data":"a01d9d82b3cfea30bfced4f72a1e28c6089568366cfb2d31ad682c7a39cec486"} Mar 21 08:48:33 crc kubenswrapper[4696]: I0321 08:48:33.385257 4696 generic.go:334] "Generic (PLEG): container finished" podID="092019b5-fac9-448e-be38-7b1cbc62f107" containerID="2e2b074e07223bf1072106b86484d2419aec5ad3e84801171bcc9236fe9e7473" exitCode=0 Mar 21 08:48:33 crc kubenswrapper[4696]: I0321 08:48:33.385375 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-p6m2c" event={"ID":"092019b5-fac9-448e-be38-7b1cbc62f107","Type":"ContainerDied","Data":"2e2b074e07223bf1072106b86484d2419aec5ad3e84801171bcc9236fe9e7473"} Mar 21 08:48:33 crc kubenswrapper[4696]: I0321 08:48:33.385396 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-p6m2c" event={"ID":"092019b5-fac9-448e-be38-7b1cbc62f107","Type":"ContainerStarted","Data":"54b3a037b1287a930b1eca507981db49098f0cd20f724e8d864b092794780639"} Mar 21 08:48:33 crc kubenswrapper[4696]: I0321 08:48:33.388111 4696 generic.go:334] "Generic (PLEG): container finished" podID="61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2" containerID="b3d0f17f9ae93f081735e0b7f189d1fc592d83b4137d6f0771c485ed6c35106d" exitCode=0 Mar 21 08:48:33 crc kubenswrapper[4696]: I0321 08:48:33.388153 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-297e-account-create-update-9rqlh" event={"ID":"61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2","Type":"ContainerDied","Data":"b3d0f17f9ae93f081735e0b7f189d1fc592d83b4137d6f0771c485ed6c35106d"} Mar 21 08:48:33 crc kubenswrapper[4696]: I0321 08:48:33.391737 4696 generic.go:334] "Generic (PLEG): container finished" podID="54341261-11d8-407a-a339-6c0d70c1ad5f" containerID="f11f6249b47a767958fabaa0043b96567dd8ae1f79f4f4bb03f10526f851234f" exitCode=0 Mar 21 08:48:33 crc kubenswrapper[4696]: I0321 08:48:33.391771 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9b07-account-create-update-g6ddw" event={"ID":"54341261-11d8-407a-a339-6c0d70c1ad5f","Type":"ContainerDied","Data":"f11f6249b47a767958fabaa0043b96567dd8ae1f79f4f4bb03f10526f851234f"} Mar 21 08:48:33 crc kubenswrapper[4696]: I0321 08:48:33.391787 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9b07-account-create-update-g6ddw" event={"ID":"54341261-11d8-407a-a339-6c0d70c1ad5f","Type":"ContainerStarted","Data":"b11238aecda45f082b2bbdd03dfbd436fa911a3733b55054ffe8af665b4b6c7a"} Mar 21 08:48:33 crc kubenswrapper[4696]: I0321 08:48:33.663872 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 21 08:48:34 crc kubenswrapper[4696]: I0321 08:48:34.413189 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1195f5f9-d86b-4038-8ce9-1573e13ed98e","Type":"ContainerStarted","Data":"3b6416f116e4a17a6ec9ab18a1f15fd44204b68a9ce729aec0f55fac552ca246"} Mar 21 08:48:34 crc kubenswrapper[4696]: I0321 08:48:34.563863 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a32ac807-7955-425a-a971-e07ddde36bb0" path="/var/lib/kubelet/pods/a32ac807-7955-425a-a971-e07ddde36bb0/volumes" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.005338 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9b07-account-create-update-g6ddw" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.107296 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54341261-11d8-407a-a339-6c0d70c1ad5f-operator-scripts\") pod \"54341261-11d8-407a-a339-6c0d70c1ad5f\" (UID: \"54341261-11d8-407a-a339-6c0d70c1ad5f\") " Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.107346 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b58tp\" (UniqueName: \"kubernetes.io/projected/54341261-11d8-407a-a339-6c0d70c1ad5f-kube-api-access-b58tp\") pod \"54341261-11d8-407a-a339-6c0d70c1ad5f\" (UID: \"54341261-11d8-407a-a339-6c0d70c1ad5f\") " Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.108658 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54341261-11d8-407a-a339-6c0d70c1ad5f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "54341261-11d8-407a-a339-6c0d70c1ad5f" (UID: "54341261-11d8-407a-a339-6c0d70c1ad5f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.128069 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54341261-11d8-407a-a339-6c0d70c1ad5f-kube-api-access-b58tp" (OuterVolumeSpecName: "kube-api-access-b58tp") pod "54341261-11d8-407a-a339-6c0d70c1ad5f" (UID: "54341261-11d8-407a-a339-6c0d70c1ad5f"). InnerVolumeSpecName "kube-api-access-b58tp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.209992 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54341261-11d8-407a-a339-6c0d70c1ad5f-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.210377 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b58tp\" (UniqueName: \"kubernetes.io/projected/54341261-11d8-407a-a339-6c0d70c1ad5f-kube-api-access-b58tp\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.313966 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-297e-account-create-update-9rqlh" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.410795 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4t9zv" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.412629 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxjg5\" (UniqueName: \"kubernetes.io/projected/61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2-kube-api-access-wxjg5\") pod \"61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2\" (UID: \"61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2\") " Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.412730 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2-operator-scripts\") pod \"61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2\" (UID: \"61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2\") " Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.413567 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2" (UID: "61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.419519 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2-kube-api-access-wxjg5" (OuterVolumeSpecName: "kube-api-access-wxjg5") pod "61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2" (UID: "61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2"). InnerVolumeSpecName "kube-api-access-wxjg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.426986 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-p6m2c" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.427697 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9172-account-create-update-r4fhw" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.427810 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8w2b4" event={"ID":"2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d","Type":"ContainerDied","Data":"e937813d81562d807962c413afb6ad61a6e55b676a7efcfae00000415e94507f"} Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.427860 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e937813d81562d807962c413afb6ad61a6e55b676a7efcfae00000415e94507f" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.433753 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9172-account-create-update-r4fhw" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.433753 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9172-account-create-update-r4fhw" event={"ID":"b9fe5e80-cc65-4225-9b74-73bf39a0f187","Type":"ContainerDied","Data":"d7d6c6438f4c1c2db6139096858304266a0d12ef9a8863f4abc0714c6ff65493"} Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.434157 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7d6c6438f4c1c2db6139096858304266a0d12ef9a8863f4abc0714c6ff65493" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.435675 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8w2b4" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.436087 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1195f5f9-d86b-4038-8ce9-1573e13ed98e","Type":"ContainerStarted","Data":"608296c55f833ffe492528423904e9c7178aa06759291b0e60a585219ca6a030"} Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.437273 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4t9zv" event={"ID":"582c3a4a-774e-4748-b055-6e2dcf1065af","Type":"ContainerDied","Data":"0d20eac6aa5e4b57ef71b33daf3a22a7704daf6292b38b8abda0274d8ba92e92"} Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.437301 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d20eac6aa5e4b57ef71b33daf3a22a7704daf6292b38b8abda0274d8ba92e92" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.437340 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4t9zv" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.438925 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-p6m2c" event={"ID":"092019b5-fac9-448e-be38-7b1cbc62f107","Type":"ContainerDied","Data":"54b3a037b1287a930b1eca507981db49098f0cd20f724e8d864b092794780639"} Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.438946 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54b3a037b1287a930b1eca507981db49098f0cd20f724e8d864b092794780639" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.438979 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-p6m2c" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.440520 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-297e-account-create-update-9rqlh" event={"ID":"61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2","Type":"ContainerDied","Data":"72108c518b505ef09c4229ffcd256bc4ffa32b056cebbf42ee01277e7d906211"} Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.440558 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72108c518b505ef09c4229ffcd256bc4ffa32b056cebbf42ee01277e7d906211" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.440612 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-297e-account-create-update-9rqlh" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.454444 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9b07-account-create-update-g6ddw" event={"ID":"54341261-11d8-407a-a339-6c0d70c1ad5f","Type":"ContainerDied","Data":"b11238aecda45f082b2bbdd03dfbd436fa911a3733b55054ffe8af665b4b6c7a"} Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.454476 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b11238aecda45f082b2bbdd03dfbd436fa911a3733b55054ffe8af665b4b6c7a" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.454568 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9b07-account-create-update-g6ddw" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.514987 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c45wf\" (UniqueName: \"kubernetes.io/projected/b9fe5e80-cc65-4225-9b74-73bf39a0f187-kube-api-access-c45wf\") pod \"b9fe5e80-cc65-4225-9b74-73bf39a0f187\" (UID: \"b9fe5e80-cc65-4225-9b74-73bf39a0f187\") " Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.515058 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/092019b5-fac9-448e-be38-7b1cbc62f107-operator-scripts\") pod \"092019b5-fac9-448e-be38-7b1cbc62f107\" (UID: \"092019b5-fac9-448e-be38-7b1cbc62f107\") " Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.515089 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrqs8\" (UniqueName: \"kubernetes.io/projected/092019b5-fac9-448e-be38-7b1cbc62f107-kube-api-access-nrqs8\") pod \"092019b5-fac9-448e-be38-7b1cbc62f107\" (UID: \"092019b5-fac9-448e-be38-7b1cbc62f107\") " Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.515114 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d-operator-scripts\") pod \"2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d\" (UID: \"2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d\") " Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.515169 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9fe5e80-cc65-4225-9b74-73bf39a0f187-operator-scripts\") pod \"b9fe5e80-cc65-4225-9b74-73bf39a0f187\" (UID: \"b9fe5e80-cc65-4225-9b74-73bf39a0f187\") " Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.515210 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgqqt\" (UniqueName: \"kubernetes.io/projected/582c3a4a-774e-4748-b055-6e2dcf1065af-kube-api-access-vgqqt\") pod \"582c3a4a-774e-4748-b055-6e2dcf1065af\" (UID: \"582c3a4a-774e-4748-b055-6e2dcf1065af\") " Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.515296 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/582c3a4a-774e-4748-b055-6e2dcf1065af-operator-scripts\") pod \"582c3a4a-774e-4748-b055-6e2dcf1065af\" (UID: \"582c3a4a-774e-4748-b055-6e2dcf1065af\") " Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.515401 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cpss\" (UniqueName: \"kubernetes.io/projected/2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d-kube-api-access-4cpss\") pod \"2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d\" (UID: \"2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d\") " Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.515808 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxjg5\" (UniqueName: \"kubernetes.io/projected/61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2-kube-api-access-wxjg5\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.515900 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.515804 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d" (UID: "2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.516094 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/092019b5-fac9-448e-be38-7b1cbc62f107-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "092019b5-fac9-448e-be38-7b1cbc62f107" (UID: "092019b5-fac9-448e-be38-7b1cbc62f107"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.516344 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9fe5e80-cc65-4225-9b74-73bf39a0f187-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b9fe5e80-cc65-4225-9b74-73bf39a0f187" (UID: "b9fe5e80-cc65-4225-9b74-73bf39a0f187"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.517851 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/582c3a4a-774e-4748-b055-6e2dcf1065af-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "582c3a4a-774e-4748-b055-6e2dcf1065af" (UID: "582c3a4a-774e-4748-b055-6e2dcf1065af"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.523481 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/582c3a4a-774e-4748-b055-6e2dcf1065af-kube-api-access-vgqqt" (OuterVolumeSpecName: "kube-api-access-vgqqt") pod "582c3a4a-774e-4748-b055-6e2dcf1065af" (UID: "582c3a4a-774e-4748-b055-6e2dcf1065af"). InnerVolumeSpecName "kube-api-access-vgqqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.523676 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9fe5e80-cc65-4225-9b74-73bf39a0f187-kube-api-access-c45wf" (OuterVolumeSpecName: "kube-api-access-c45wf") pod "b9fe5e80-cc65-4225-9b74-73bf39a0f187" (UID: "b9fe5e80-cc65-4225-9b74-73bf39a0f187"). InnerVolumeSpecName "kube-api-access-c45wf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.526307 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/092019b5-fac9-448e-be38-7b1cbc62f107-kube-api-access-nrqs8" (OuterVolumeSpecName: "kube-api-access-nrqs8") pod "092019b5-fac9-448e-be38-7b1cbc62f107" (UID: "092019b5-fac9-448e-be38-7b1cbc62f107"). InnerVolumeSpecName "kube-api-access-nrqs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.526964 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d-kube-api-access-4cpss" (OuterVolumeSpecName: "kube-api-access-4cpss") pod "2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d" (UID: "2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d"). InnerVolumeSpecName "kube-api-access-4cpss". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.617179 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cpss\" (UniqueName: \"kubernetes.io/projected/2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d-kube-api-access-4cpss\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.617227 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c45wf\" (UniqueName: \"kubernetes.io/projected/b9fe5e80-cc65-4225-9b74-73bf39a0f187-kube-api-access-c45wf\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.617239 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/092019b5-fac9-448e-be38-7b1cbc62f107-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.617249 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrqs8\" (UniqueName: \"kubernetes.io/projected/092019b5-fac9-448e-be38-7b1cbc62f107-kube-api-access-nrqs8\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.617257 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.617265 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9fe5e80-cc65-4225-9b74-73bf39a0f187-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.617274 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgqqt\" (UniqueName: \"kubernetes.io/projected/582c3a4a-774e-4748-b055-6e2dcf1065af-kube-api-access-vgqqt\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:35 crc kubenswrapper[4696]: I0321 08:48:35.617282 4696 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/582c3a4a-774e-4748-b055-6e2dcf1065af-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:35 crc kubenswrapper[4696]: E0321 08:48:35.853279 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod092019b5_fac9_448e_be38_7b1cbc62f107.slice/crio-54b3a037b1287a930b1eca507981db49098f0cd20f724e8d864b092794780639\": RecentStats: unable to find data in memory cache]" Mar 21 08:48:36 crc kubenswrapper[4696]: I0321 08:48:36.481351 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1195f5f9-d86b-4038-8ce9-1573e13ed98e","Type":"ContainerStarted","Data":"e0897fd115535748ef0cb76530ca8c0943d0ee3ec9599ced057539820dee69dd"} Mar 21 08:48:36 crc kubenswrapper[4696]: I0321 08:48:36.481407 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8w2b4" Mar 21 08:48:37 crc kubenswrapper[4696]: I0321 08:48:37.712855 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:37 crc kubenswrapper[4696]: I0321 08:48:37.713863 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7b65745bbc-5b5zz" Mar 21 08:48:37 crc kubenswrapper[4696]: I0321 08:48:37.957263 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-69b5b45b6f-7lhn7" Mar 21 08:48:38 crc kubenswrapper[4696]: I0321 08:48:38.068316 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5d77f68d84-8hrbc"] Mar 21 08:48:38 crc kubenswrapper[4696]: I0321 08:48:38.068538 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5d77f68d84-8hrbc" podUID="9cba9c3e-f501-4351-b0af-26c07ff8c757" containerName="neutron-api" containerID="cri-o://fe5af300da48a9658e1669e9924993a9cbf0038e20390138a155bf19acc5456a" gracePeriod=30 Mar 21 08:48:38 crc kubenswrapper[4696]: I0321 08:48:38.068941 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5d77f68d84-8hrbc" podUID="9cba9c3e-f501-4351-b0af-26c07ff8c757" containerName="neutron-httpd" containerID="cri-o://cb493038d7980c2cefcca5b04d8e6520ea16a047f9b766eef873b8ed1ece417d" gracePeriod=30 Mar 21 08:48:38 crc kubenswrapper[4696]: I0321 08:48:38.414537 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:48:38 crc kubenswrapper[4696]: I0321 08:48:38.514243 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1195f5f9-d86b-4038-8ce9-1573e13ed98e","Type":"ContainerStarted","Data":"8197c0636f36b4aed7821f41955fae8ff0c1ec9a613e22ba3586a01a7f40f1e7"} Mar 21 08:48:38 crc kubenswrapper[4696]: I0321 08:48:38.515428 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 21 08:48:38 crc kubenswrapper[4696]: I0321 08:48:38.526229 4696 generic.go:334] "Generic (PLEG): container finished" podID="9cba9c3e-f501-4351-b0af-26c07ff8c757" containerID="cb493038d7980c2cefcca5b04d8e6520ea16a047f9b766eef873b8ed1ece417d" exitCode=0 Mar 21 08:48:38 crc kubenswrapper[4696]: I0321 08:48:38.527097 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d77f68d84-8hrbc" event={"ID":"9cba9c3e-f501-4351-b0af-26c07ff8c757","Type":"ContainerDied","Data":"cb493038d7980c2cefcca5b04d8e6520ea16a047f9b766eef873b8ed1ece417d"} Mar 21 08:48:38 crc kubenswrapper[4696]: I0321 08:48:38.532976 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.162465264 podStartE2EDuration="7.532961763s" podCreationTimestamp="2026-03-21 08:48:31 +0000 UTC" firstStartedPulling="2026-03-21 08:48:32.580810846 +0000 UTC m=+1246.701691559" lastFinishedPulling="2026-03-21 08:48:37.951307345 +0000 UTC m=+1252.072188058" observedRunningTime="2026-03-21 08:48:38.532252223 +0000 UTC m=+1252.653132936" watchObservedRunningTime="2026-03-21 08:48:38.532961763 +0000 UTC m=+1252.653842476" Mar 21 08:48:39 crc kubenswrapper[4696]: I0321 08:48:39.545840 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1195f5f9-d86b-4038-8ce9-1573e13ed98e" containerName="ceilometer-central-agent" containerID="cri-o://3b6416f116e4a17a6ec9ab18a1f15fd44204b68a9ce729aec0f55fac552ca246" gracePeriod=30 Mar 21 08:48:39 crc kubenswrapper[4696]: I0321 08:48:39.546604 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1195f5f9-d86b-4038-8ce9-1573e13ed98e" containerName="proxy-httpd" containerID="cri-o://8197c0636f36b4aed7821f41955fae8ff0c1ec9a613e22ba3586a01a7f40f1e7" gracePeriod=30 Mar 21 08:48:39 crc kubenswrapper[4696]: I0321 08:48:39.546655 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1195f5f9-d86b-4038-8ce9-1573e13ed98e" containerName="sg-core" containerID="cri-o://e0897fd115535748ef0cb76530ca8c0943d0ee3ec9599ced057539820dee69dd" gracePeriod=30 Mar 21 08:48:39 crc kubenswrapper[4696]: I0321 08:48:39.546712 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1195f5f9-d86b-4038-8ce9-1573e13ed98e" containerName="ceilometer-notification-agent" containerID="cri-o://608296c55f833ffe492528423904e9c7178aa06759291b0e60a585219ca6a030" gracePeriod=30 Mar 21 08:48:40 crc kubenswrapper[4696]: I0321 08:48:40.557967 4696 generic.go:334] "Generic (PLEG): container finished" podID="1195f5f9-d86b-4038-8ce9-1573e13ed98e" containerID="8197c0636f36b4aed7821f41955fae8ff0c1ec9a613e22ba3586a01a7f40f1e7" exitCode=0 Mar 21 08:48:40 crc kubenswrapper[4696]: I0321 08:48:40.558322 4696 generic.go:334] "Generic (PLEG): container finished" podID="1195f5f9-d86b-4038-8ce9-1573e13ed98e" containerID="e0897fd115535748ef0cb76530ca8c0943d0ee3ec9599ced057539820dee69dd" exitCode=2 Mar 21 08:48:40 crc kubenswrapper[4696]: I0321 08:48:40.558331 4696 generic.go:334] "Generic (PLEG): container finished" podID="1195f5f9-d86b-4038-8ce9-1573e13ed98e" containerID="608296c55f833ffe492528423904e9c7178aa06759291b0e60a585219ca6a030" exitCode=0 Mar 21 08:48:40 crc kubenswrapper[4696]: I0321 08:48:40.558344 4696 generic.go:334] "Generic (PLEG): container finished" podID="1195f5f9-d86b-4038-8ce9-1573e13ed98e" containerID="3b6416f116e4a17a6ec9ab18a1f15fd44204b68a9ce729aec0f55fac552ca246" exitCode=0 Mar 21 08:48:40 crc kubenswrapper[4696]: I0321 08:48:40.558068 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1195f5f9-d86b-4038-8ce9-1573e13ed98e","Type":"ContainerDied","Data":"8197c0636f36b4aed7821f41955fae8ff0c1ec9a613e22ba3586a01a7f40f1e7"} Mar 21 08:48:40 crc kubenswrapper[4696]: I0321 08:48:40.558378 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1195f5f9-d86b-4038-8ce9-1573e13ed98e","Type":"ContainerDied","Data":"e0897fd115535748ef0cb76530ca8c0943d0ee3ec9599ced057539820dee69dd"} Mar 21 08:48:40 crc kubenswrapper[4696]: I0321 08:48:40.558394 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1195f5f9-d86b-4038-8ce9-1573e13ed98e","Type":"ContainerDied","Data":"608296c55f833ffe492528423904e9c7178aa06759291b0e60a585219ca6a030"} Mar 21 08:48:40 crc kubenswrapper[4696]: I0321 08:48:40.558405 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1195f5f9-d86b-4038-8ce9-1573e13ed98e","Type":"ContainerDied","Data":"3b6416f116e4a17a6ec9ab18a1f15fd44204b68a9ce729aec0f55fac552ca246"} Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.133675 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-j298g"] Mar 21 08:48:41 crc kubenswrapper[4696]: E0321 08:48:41.134351 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="582c3a4a-774e-4748-b055-6e2dcf1065af" containerName="mariadb-database-create" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.134369 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="582c3a4a-774e-4748-b055-6e2dcf1065af" containerName="mariadb-database-create" Mar 21 08:48:41 crc kubenswrapper[4696]: E0321 08:48:41.134382 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a32ac807-7955-425a-a971-e07ddde36bb0" containerName="placement-api" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.134388 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="a32ac807-7955-425a-a971-e07ddde36bb0" containerName="placement-api" Mar 21 08:48:41 crc kubenswrapper[4696]: E0321 08:48:41.134402 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2" containerName="mariadb-account-create-update" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.134408 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2" containerName="mariadb-account-create-update" Mar 21 08:48:41 crc kubenswrapper[4696]: E0321 08:48:41.134431 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d" containerName="mariadb-database-create" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.134437 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d" containerName="mariadb-database-create" Mar 21 08:48:41 crc kubenswrapper[4696]: E0321 08:48:41.134448 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54341261-11d8-407a-a339-6c0d70c1ad5f" containerName="mariadb-account-create-update" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.134454 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="54341261-11d8-407a-a339-6c0d70c1ad5f" containerName="mariadb-account-create-update" Mar 21 08:48:41 crc kubenswrapper[4696]: E0321 08:48:41.134465 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="092019b5-fac9-448e-be38-7b1cbc62f107" containerName="mariadb-database-create" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.134471 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="092019b5-fac9-448e-be38-7b1cbc62f107" containerName="mariadb-database-create" Mar 21 08:48:41 crc kubenswrapper[4696]: E0321 08:48:41.134484 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a32ac807-7955-425a-a971-e07ddde36bb0" containerName="placement-log" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.134491 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="a32ac807-7955-425a-a971-e07ddde36bb0" containerName="placement-log" Mar 21 08:48:41 crc kubenswrapper[4696]: E0321 08:48:41.134509 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9fe5e80-cc65-4225-9b74-73bf39a0f187" containerName="mariadb-account-create-update" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.134516 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9fe5e80-cc65-4225-9b74-73bf39a0f187" containerName="mariadb-account-create-update" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.134682 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="582c3a4a-774e-4748-b055-6e2dcf1065af" containerName="mariadb-database-create" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.134696 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="092019b5-fac9-448e-be38-7b1cbc62f107" containerName="mariadb-database-create" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.134704 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9fe5e80-cc65-4225-9b74-73bf39a0f187" containerName="mariadb-account-create-update" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.134716 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="a32ac807-7955-425a-a971-e07ddde36bb0" containerName="placement-api" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.134726 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2" containerName="mariadb-account-create-update" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.134734 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="54341261-11d8-407a-a339-6c0d70c1ad5f" containerName="mariadb-account-create-update" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.134746 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="a32ac807-7955-425a-a971-e07ddde36bb0" containerName="placement-log" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.134756 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d" containerName="mariadb-database-create" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.135422 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-j298g" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.138141 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.138209 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-psnkw" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.143856 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.152996 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-j298g"] Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.247634 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m58v5\" (UniqueName: \"kubernetes.io/projected/5e448486-5596-4442-a11a-804a7b7b94d8-kube-api-access-m58v5\") pod \"nova-cell0-conductor-db-sync-j298g\" (UID: \"5e448486-5596-4442-a11a-804a7b7b94d8\") " pod="openstack/nova-cell0-conductor-db-sync-j298g" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.247764 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e448486-5596-4442-a11a-804a7b7b94d8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-j298g\" (UID: \"5e448486-5596-4442-a11a-804a7b7b94d8\") " pod="openstack/nova-cell0-conductor-db-sync-j298g" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.247836 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e448486-5596-4442-a11a-804a7b7b94d8-config-data\") pod \"nova-cell0-conductor-db-sync-j298g\" (UID: \"5e448486-5596-4442-a11a-804a7b7b94d8\") " pod="openstack/nova-cell0-conductor-db-sync-j298g" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.248149 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e448486-5596-4442-a11a-804a7b7b94d8-scripts\") pod \"nova-cell0-conductor-db-sync-j298g\" (UID: \"5e448486-5596-4442-a11a-804a7b7b94d8\") " pod="openstack/nova-cell0-conductor-db-sync-j298g" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.350242 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e448486-5596-4442-a11a-804a7b7b94d8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-j298g\" (UID: \"5e448486-5596-4442-a11a-804a7b7b94d8\") " pod="openstack/nova-cell0-conductor-db-sync-j298g" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.350333 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e448486-5596-4442-a11a-804a7b7b94d8-config-data\") pod \"nova-cell0-conductor-db-sync-j298g\" (UID: \"5e448486-5596-4442-a11a-804a7b7b94d8\") " pod="openstack/nova-cell0-conductor-db-sync-j298g" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.350421 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e448486-5596-4442-a11a-804a7b7b94d8-scripts\") pod \"nova-cell0-conductor-db-sync-j298g\" (UID: \"5e448486-5596-4442-a11a-804a7b7b94d8\") " pod="openstack/nova-cell0-conductor-db-sync-j298g" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.350464 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m58v5\" (UniqueName: \"kubernetes.io/projected/5e448486-5596-4442-a11a-804a7b7b94d8-kube-api-access-m58v5\") pod \"nova-cell0-conductor-db-sync-j298g\" (UID: \"5e448486-5596-4442-a11a-804a7b7b94d8\") " pod="openstack/nova-cell0-conductor-db-sync-j298g" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.356249 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e448486-5596-4442-a11a-804a7b7b94d8-scripts\") pod \"nova-cell0-conductor-db-sync-j298g\" (UID: \"5e448486-5596-4442-a11a-804a7b7b94d8\") " pod="openstack/nova-cell0-conductor-db-sync-j298g" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.360390 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e448486-5596-4442-a11a-804a7b7b94d8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-j298g\" (UID: \"5e448486-5596-4442-a11a-804a7b7b94d8\") " pod="openstack/nova-cell0-conductor-db-sync-j298g" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.361085 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e448486-5596-4442-a11a-804a7b7b94d8-config-data\") pod \"nova-cell0-conductor-db-sync-j298g\" (UID: \"5e448486-5596-4442-a11a-804a7b7b94d8\") " pod="openstack/nova-cell0-conductor-db-sync-j298g" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.373344 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m58v5\" (UniqueName: \"kubernetes.io/projected/5e448486-5596-4442-a11a-804a7b7b94d8-kube-api-access-m58v5\") pod \"nova-cell0-conductor-db-sync-j298g\" (UID: \"5e448486-5596-4442-a11a-804a7b7b94d8\") " pod="openstack/nova-cell0-conductor-db-sync-j298g" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.456776 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-j298g" Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.572661 4696 generic.go:334] "Generic (PLEG): container finished" podID="9cba9c3e-f501-4351-b0af-26c07ff8c757" containerID="fe5af300da48a9658e1669e9924993a9cbf0038e20390138a155bf19acc5456a" exitCode=0 Mar 21 08:48:41 crc kubenswrapper[4696]: I0321 08:48:41.572701 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d77f68d84-8hrbc" event={"ID":"9cba9c3e-f501-4351-b0af-26c07ff8c757","Type":"ContainerDied","Data":"fe5af300da48a9658e1669e9924993a9cbf0038e20390138a155bf19acc5456a"} Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.418481 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d77f68d84-8hrbc" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.471701 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.541216 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-scripts\") pod \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.541282 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1195f5f9-d86b-4038-8ce9-1573e13ed98e-run-httpd\") pod \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.541341 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xml8b\" (UniqueName: \"kubernetes.io/projected/1195f5f9-d86b-4038-8ce9-1573e13ed98e-kube-api-access-xml8b\") pod \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.541369 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-config-data\") pod \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.541449 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-combined-ca-bundle\") pod \"9cba9c3e-f501-4351-b0af-26c07ff8c757\" (UID: \"9cba9c3e-f501-4351-b0af-26c07ff8c757\") " Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.541487 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1195f5f9-d86b-4038-8ce9-1573e13ed98e-log-httpd\") pod \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.541535 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-ovndb-tls-certs\") pod \"9cba9c3e-f501-4351-b0af-26c07ff8c757\" (UID: \"9cba9c3e-f501-4351-b0af-26c07ff8c757\") " Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.541566 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-httpd-config\") pod \"9cba9c3e-f501-4351-b0af-26c07ff8c757\" (UID: \"9cba9c3e-f501-4351-b0af-26c07ff8c757\") " Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.541586 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-sg-core-conf-yaml\") pod \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.541611 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwfxf\" (UniqueName: \"kubernetes.io/projected/9cba9c3e-f501-4351-b0af-26c07ff8c757-kube-api-access-vwfxf\") pod \"9cba9c3e-f501-4351-b0af-26c07ff8c757\" (UID: \"9cba9c3e-f501-4351-b0af-26c07ff8c757\") " Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.541633 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-config\") pod \"9cba9c3e-f501-4351-b0af-26c07ff8c757\" (UID: \"9cba9c3e-f501-4351-b0af-26c07ff8c757\") " Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.541683 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-combined-ca-bundle\") pod \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\" (UID: \"1195f5f9-d86b-4038-8ce9-1573e13ed98e\") " Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.541811 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1195f5f9-d86b-4038-8ce9-1573e13ed98e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1195f5f9-d86b-4038-8ce9-1573e13ed98e" (UID: "1195f5f9-d86b-4038-8ce9-1573e13ed98e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.542119 4696 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1195f5f9-d86b-4038-8ce9-1573e13ed98e-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.542371 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1195f5f9-d86b-4038-8ce9-1573e13ed98e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1195f5f9-d86b-4038-8ce9-1573e13ed98e" (UID: "1195f5f9-d86b-4038-8ce9-1573e13ed98e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.557158 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-scripts" (OuterVolumeSpecName: "scripts") pod "1195f5f9-d86b-4038-8ce9-1573e13ed98e" (UID: "1195f5f9-d86b-4038-8ce9-1573e13ed98e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.557166 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1195f5f9-d86b-4038-8ce9-1573e13ed98e-kube-api-access-xml8b" (OuterVolumeSpecName: "kube-api-access-xml8b") pod "1195f5f9-d86b-4038-8ce9-1573e13ed98e" (UID: "1195f5f9-d86b-4038-8ce9-1573e13ed98e"). InnerVolumeSpecName "kube-api-access-xml8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.557152 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cba9c3e-f501-4351-b0af-26c07ff8c757-kube-api-access-vwfxf" (OuterVolumeSpecName: "kube-api-access-vwfxf") pod "9cba9c3e-f501-4351-b0af-26c07ff8c757" (UID: "9cba9c3e-f501-4351-b0af-26c07ff8c757"). InnerVolumeSpecName "kube-api-access-vwfxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.557227 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "9cba9c3e-f501-4351-b0af-26c07ff8c757" (UID: "9cba9c3e-f501-4351-b0af-26c07ff8c757"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.598240 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1195f5f9-d86b-4038-8ce9-1573e13ed98e" (UID: "1195f5f9-d86b-4038-8ce9-1573e13ed98e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.612494 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d77f68d84-8hrbc" event={"ID":"9cba9c3e-f501-4351-b0af-26c07ff8c757","Type":"ContainerDied","Data":"540d30b2c60a2e402249c23a1cbbbcd811f4ad5d55f8b474e447c905ac1eb40f"} Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.614012 4696 scope.go:117] "RemoveContainer" containerID="cb493038d7980c2cefcca5b04d8e6520ea16a047f9b766eef873b8ed1ece417d" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.612699 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d77f68d84-8hrbc" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.617489 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1195f5f9-d86b-4038-8ce9-1573e13ed98e","Type":"ContainerDied","Data":"7fe3c8481a51d7859bef649640eb629a0eb5cae6a143aacf8f505e978df280ed"} Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.617617 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.619098 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-j298g"] Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.624117 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"383967e9-132f-4355-abbb-5904830137b7","Type":"ContainerStarted","Data":"435a7237ec1d046b632bc186abf2064f4a38818f81b4e962171bbed40703a648"} Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.638276 4696 scope.go:117] "RemoveContainer" containerID="fe5af300da48a9658e1669e9924993a9cbf0038e20390138a155bf19acc5456a" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.642120 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-config" (OuterVolumeSpecName: "config") pod "9cba9c3e-f501-4351-b0af-26c07ff8c757" (UID: "9cba9c3e-f501-4351-b0af-26c07ff8c757"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.644182 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.644205 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xml8b\" (UniqueName: \"kubernetes.io/projected/1195f5f9-d86b-4038-8ce9-1573e13ed98e-kube-api-access-xml8b\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.644215 4696 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1195f5f9-d86b-4038-8ce9-1573e13ed98e-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.644226 4696 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.644238 4696 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.646316 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwfxf\" (UniqueName: \"kubernetes.io/projected/9cba9c3e-f501-4351-b0af-26c07ff8c757-kube-api-access-vwfxf\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.646340 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.658973 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1195f5f9-d86b-4038-8ce9-1573e13ed98e" (UID: "1195f5f9-d86b-4038-8ce9-1573e13ed98e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.662017 4696 scope.go:117] "RemoveContainer" containerID="8197c0636f36b4aed7821f41955fae8ff0c1ec9a613e22ba3586a01a7f40f1e7" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.662190 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.261722441 podStartE2EDuration="20.662168303s" podCreationTimestamp="2026-03-21 08:48:25 +0000 UTC" firstStartedPulling="2026-03-21 08:48:26.500589994 +0000 UTC m=+1240.621470707" lastFinishedPulling="2026-03-21 08:48:44.901035856 +0000 UTC m=+1259.021916569" observedRunningTime="2026-03-21 08:48:45.651431626 +0000 UTC m=+1259.772312359" watchObservedRunningTime="2026-03-21 08:48:45.662168303 +0000 UTC m=+1259.783049016" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.674021 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9cba9c3e-f501-4351-b0af-26c07ff8c757" (UID: "9cba9c3e-f501-4351-b0af-26c07ff8c757"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.676557 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "9cba9c3e-f501-4351-b0af-26c07ff8c757" (UID: "9cba9c3e-f501-4351-b0af-26c07ff8c757"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.680279 4696 scope.go:117] "RemoveContainer" containerID="e0897fd115535748ef0cb76530ca8c0943d0ee3ec9599ced057539820dee69dd" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.701839 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-config-data" (OuterVolumeSpecName: "config-data") pod "1195f5f9-d86b-4038-8ce9-1573e13ed98e" (UID: "1195f5f9-d86b-4038-8ce9-1573e13ed98e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.703010 4696 scope.go:117] "RemoveContainer" containerID="608296c55f833ffe492528423904e9c7178aa06759291b0e60a585219ca6a030" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.724804 4696 scope.go:117] "RemoveContainer" containerID="3b6416f116e4a17a6ec9ab18a1f15fd44204b68a9ce729aec0f55fac552ca246" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.748606 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.748641 4696 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cba9c3e-f501-4351-b0af-26c07ff8c757-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.748655 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.748676 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1195f5f9-d86b-4038-8ce9-1573e13ed98e-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.964432 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.978002 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:48:45 crc kubenswrapper[4696]: I0321 08:48:45.991055 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5d77f68d84-8hrbc"] Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.009628 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5d77f68d84-8hrbc"] Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.019327 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:48:46 crc kubenswrapper[4696]: E0321 08:48:46.019903 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cba9c3e-f501-4351-b0af-26c07ff8c757" containerName="neutron-httpd" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.020002 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cba9c3e-f501-4351-b0af-26c07ff8c757" containerName="neutron-httpd" Mar 21 08:48:46 crc kubenswrapper[4696]: E0321 08:48:46.020094 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1195f5f9-d86b-4038-8ce9-1573e13ed98e" containerName="proxy-httpd" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.020148 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1195f5f9-d86b-4038-8ce9-1573e13ed98e" containerName="proxy-httpd" Mar 21 08:48:46 crc kubenswrapper[4696]: E0321 08:48:46.020224 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1195f5f9-d86b-4038-8ce9-1573e13ed98e" containerName="sg-core" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.020283 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1195f5f9-d86b-4038-8ce9-1573e13ed98e" containerName="sg-core" Mar 21 08:48:46 crc kubenswrapper[4696]: E0321 08:48:46.020338 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1195f5f9-d86b-4038-8ce9-1573e13ed98e" containerName="ceilometer-notification-agent" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.020394 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1195f5f9-d86b-4038-8ce9-1573e13ed98e" containerName="ceilometer-notification-agent" Mar 21 08:48:46 crc kubenswrapper[4696]: E0321 08:48:46.020467 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1195f5f9-d86b-4038-8ce9-1573e13ed98e" containerName="ceilometer-central-agent" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.020520 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1195f5f9-d86b-4038-8ce9-1573e13ed98e" containerName="ceilometer-central-agent" Mar 21 08:48:46 crc kubenswrapper[4696]: E0321 08:48:46.020589 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cba9c3e-f501-4351-b0af-26c07ff8c757" containerName="neutron-api" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.020646 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cba9c3e-f501-4351-b0af-26c07ff8c757" containerName="neutron-api" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.020879 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1195f5f9-d86b-4038-8ce9-1573e13ed98e" containerName="ceilometer-central-agent" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.020953 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cba9c3e-f501-4351-b0af-26c07ff8c757" containerName="neutron-httpd" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.021031 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cba9c3e-f501-4351-b0af-26c07ff8c757" containerName="neutron-api" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.021094 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1195f5f9-d86b-4038-8ce9-1573e13ed98e" containerName="proxy-httpd" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.021150 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1195f5f9-d86b-4038-8ce9-1573e13ed98e" containerName="ceilometer-notification-agent" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.021206 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1195f5f9-d86b-4038-8ce9-1573e13ed98e" containerName="sg-core" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.023080 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.025368 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.028426 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.028741 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.052806 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhg6k\" (UniqueName: \"kubernetes.io/projected/f273604b-ebe8-4818-8d9c-322bd8dcfe93-kube-api-access-lhg6k\") pod \"ceilometer-0\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.053078 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-scripts\") pod \"ceilometer-0\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.053205 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.053330 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f273604b-ebe8-4818-8d9c-322bd8dcfe93-run-httpd\") pod \"ceilometer-0\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.053417 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f273604b-ebe8-4818-8d9c-322bd8dcfe93-log-httpd\") pod \"ceilometer-0\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.053508 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-config-data\") pod \"ceilometer-0\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.053698 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: E0321 08:48:46.139444 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1195f5f9_d86b_4038_8ce9_1573e13ed98e.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1195f5f9_d86b_4038_8ce9_1573e13ed98e.slice/crio-7fe3c8481a51d7859bef649640eb629a0eb5cae6a143aacf8f505e978df280ed\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9cba9c3e_f501_4351_b0af_26c07ff8c757.slice\": RecentStats: unable to find data in memory cache]" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.156320 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhg6k\" (UniqueName: \"kubernetes.io/projected/f273604b-ebe8-4818-8d9c-322bd8dcfe93-kube-api-access-lhg6k\") pod \"ceilometer-0\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.156697 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-scripts\") pod \"ceilometer-0\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.156737 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.156835 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f273604b-ebe8-4818-8d9c-322bd8dcfe93-run-httpd\") pod \"ceilometer-0\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.156859 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f273604b-ebe8-4818-8d9c-322bd8dcfe93-log-httpd\") pod \"ceilometer-0\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.156893 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-config-data\") pod \"ceilometer-0\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.157206 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.157346 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f273604b-ebe8-4818-8d9c-322bd8dcfe93-log-httpd\") pod \"ceilometer-0\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.157919 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f273604b-ebe8-4818-8d9c-322bd8dcfe93-run-httpd\") pod \"ceilometer-0\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.163155 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.164269 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-scripts\") pod \"ceilometer-0\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.165450 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-config-data\") pod \"ceilometer-0\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.172648 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.173005 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhg6k\" (UniqueName: \"kubernetes.io/projected/f273604b-ebe8-4818-8d9c-322bd8dcfe93-kube-api-access-lhg6k\") pod \"ceilometer-0\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.339585 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.552741 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1195f5f9-d86b-4038-8ce9-1573e13ed98e" path="/var/lib/kubelet/pods/1195f5f9-d86b-4038-8ce9-1573e13ed98e/volumes" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.553877 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cba9c3e-f501-4351-b0af-26c07ff8c757" path="/var/lib/kubelet/pods/9cba9c3e-f501-4351-b0af-26c07ff8c757/volumes" Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.653381 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-j298g" event={"ID":"5e448486-5596-4442-a11a-804a7b7b94d8","Type":"ContainerStarted","Data":"371f9e37d3154054819d68cf79e88d5ee346e2c14faf64f6d34a041e087dc079"} Mar 21 08:48:46 crc kubenswrapper[4696]: I0321 08:48:46.813789 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:48:46 crc kubenswrapper[4696]: W0321 08:48:46.816945 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf273604b_ebe8_4818_8d9c_322bd8dcfe93.slice/crio-36a2a47ecfd009b853f3d409174c86938f21d43d270c800ee42c4255f17898ec WatchSource:0}: Error finding container 36a2a47ecfd009b853f3d409174c86938f21d43d270c800ee42c4255f17898ec: Status 404 returned error can't find the container with id 36a2a47ecfd009b853f3d409174c86938f21d43d270c800ee42c4255f17898ec Mar 21 08:48:47 crc kubenswrapper[4696]: I0321 08:48:47.682338 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f273604b-ebe8-4818-8d9c-322bd8dcfe93","Type":"ContainerStarted","Data":"302d82ab7099d32e78e67260e0cc48405091bcaaa141d5555faaa5e07d801479"} Mar 21 08:48:47 crc kubenswrapper[4696]: I0321 08:48:47.682670 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f273604b-ebe8-4818-8d9c-322bd8dcfe93","Type":"ContainerStarted","Data":"36a2a47ecfd009b853f3d409174c86938f21d43d270c800ee42c4255f17898ec"} Mar 21 08:48:48 crc kubenswrapper[4696]: I0321 08:48:48.696247 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f273604b-ebe8-4818-8d9c-322bd8dcfe93","Type":"ContainerStarted","Data":"139de8563bb4dbb9b04883c62b1ac4af298b15486793bf129d01e835f9cc2a61"} Mar 21 08:48:49 crc kubenswrapper[4696]: I0321 08:48:49.708380 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f273604b-ebe8-4818-8d9c-322bd8dcfe93","Type":"ContainerStarted","Data":"8c44d93273437f52d1efdbc780e083274abd6cc990ac7c6831b2bdd5ece0fc89"} Mar 21 08:48:50 crc kubenswrapper[4696]: I0321 08:48:50.266750 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:48:54 crc kubenswrapper[4696]: I0321 08:48:54.702684 4696 scope.go:117] "RemoveContainer" containerID="fe52e941866d9eb50ec7713e66e65174156613e23766607e0802e44303d6d7f6" Mar 21 08:48:57 crc kubenswrapper[4696]: I0321 08:48:57.070808 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-api-0" Mar 21 08:48:57 crc kubenswrapper[4696]: I0321 08:48:57.836230 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-j298g" event={"ID":"5e448486-5596-4442-a11a-804a7b7b94d8","Type":"ContainerStarted","Data":"9eae8c5cbd9fc1b2ae5e65811cb3ddc5f1961c88f5db9e6c115709c50c3f3c03"} Mar 21 08:48:57 crc kubenswrapper[4696]: I0321 08:48:57.839228 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f273604b-ebe8-4818-8d9c-322bd8dcfe93","Type":"ContainerStarted","Data":"5839a6fcfbe10b15e983b0c9e75e9258e17020dc98ee4a1dc8cc88f9c2f4d14d"} Mar 21 08:48:57 crc kubenswrapper[4696]: I0321 08:48:57.839496 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f273604b-ebe8-4818-8d9c-322bd8dcfe93" containerName="sg-core" containerID="cri-o://8c44d93273437f52d1efdbc780e083274abd6cc990ac7c6831b2bdd5ece0fc89" gracePeriod=30 Mar 21 08:48:57 crc kubenswrapper[4696]: I0321 08:48:57.839535 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f273604b-ebe8-4818-8d9c-322bd8dcfe93" containerName="ceilometer-notification-agent" containerID="cri-o://139de8563bb4dbb9b04883c62b1ac4af298b15486793bf129d01e835f9cc2a61" gracePeriod=30 Mar 21 08:48:57 crc kubenswrapper[4696]: I0321 08:48:57.839583 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f273604b-ebe8-4818-8d9c-322bd8dcfe93" containerName="proxy-httpd" containerID="cri-o://5839a6fcfbe10b15e983b0c9e75e9258e17020dc98ee4a1dc8cc88f9c2f4d14d" gracePeriod=30 Mar 21 08:48:57 crc kubenswrapper[4696]: I0321 08:48:57.839694 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 21 08:48:57 crc kubenswrapper[4696]: I0321 08:48:57.839460 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f273604b-ebe8-4818-8d9c-322bd8dcfe93" containerName="ceilometer-central-agent" containerID="cri-o://302d82ab7099d32e78e67260e0cc48405091bcaaa141d5555faaa5e07d801479" gracePeriod=30 Mar 21 08:48:57 crc kubenswrapper[4696]: I0321 08:48:57.862681 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-j298g" podStartSLOduration=5.676226644 podStartE2EDuration="16.862656745s" podCreationTimestamp="2026-03-21 08:48:41 +0000 UTC" firstStartedPulling="2026-03-21 08:48:45.611717286 +0000 UTC m=+1259.732597999" lastFinishedPulling="2026-03-21 08:48:56.798147377 +0000 UTC m=+1270.919028100" observedRunningTime="2026-03-21 08:48:57.857957165 +0000 UTC m=+1271.978837878" watchObservedRunningTime="2026-03-21 08:48:57.862656745 +0000 UTC m=+1271.983537458" Mar 21 08:48:57 crc kubenswrapper[4696]: I0321 08:48:57.893855 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.891470928 podStartE2EDuration="12.893834169s" podCreationTimestamp="2026-03-21 08:48:45 +0000 UTC" firstStartedPulling="2026-03-21 08:48:46.820101129 +0000 UTC m=+1260.940981842" lastFinishedPulling="2026-03-21 08:48:56.82246437 +0000 UTC m=+1270.943345083" observedRunningTime="2026-03-21 08:48:57.883720668 +0000 UTC m=+1272.004601381" watchObservedRunningTime="2026-03-21 08:48:57.893834169 +0000 UTC m=+1272.014714882" Mar 21 08:48:58 crc kubenswrapper[4696]: I0321 08:48:58.858137 4696 generic.go:334] "Generic (PLEG): container finished" podID="f273604b-ebe8-4818-8d9c-322bd8dcfe93" containerID="5839a6fcfbe10b15e983b0c9e75e9258e17020dc98ee4a1dc8cc88f9c2f4d14d" exitCode=0 Mar 21 08:48:58 crc kubenswrapper[4696]: I0321 08:48:58.858424 4696 generic.go:334] "Generic (PLEG): container finished" podID="f273604b-ebe8-4818-8d9c-322bd8dcfe93" containerID="8c44d93273437f52d1efdbc780e083274abd6cc990ac7c6831b2bdd5ece0fc89" exitCode=2 Mar 21 08:48:58 crc kubenswrapper[4696]: I0321 08:48:58.858432 4696 generic.go:334] "Generic (PLEG): container finished" podID="f273604b-ebe8-4818-8d9c-322bd8dcfe93" containerID="139de8563bb4dbb9b04883c62b1ac4af298b15486793bf129d01e835f9cc2a61" exitCode=0 Mar 21 08:48:58 crc kubenswrapper[4696]: I0321 08:48:58.858209 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f273604b-ebe8-4818-8d9c-322bd8dcfe93","Type":"ContainerDied","Data":"5839a6fcfbe10b15e983b0c9e75e9258e17020dc98ee4a1dc8cc88f9c2f4d14d"} Mar 21 08:48:58 crc kubenswrapper[4696]: I0321 08:48:58.858550 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f273604b-ebe8-4818-8d9c-322bd8dcfe93","Type":"ContainerDied","Data":"8c44d93273437f52d1efdbc780e083274abd6cc990ac7c6831b2bdd5ece0fc89"} Mar 21 08:48:58 crc kubenswrapper[4696]: I0321 08:48:58.858579 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f273604b-ebe8-4818-8d9c-322bd8dcfe93","Type":"ContainerDied","Data":"139de8563bb4dbb9b04883c62b1ac4af298b15486793bf129d01e835f9cc2a61"} Mar 21 08:48:58 crc kubenswrapper[4696]: I0321 08:48:58.950114 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 08:48:58 crc kubenswrapper[4696]: I0321 08:48:58.950357 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="02fca1e0-3685-4d5b-86f9-bef0511876df" containerName="glance-log" containerID="cri-o://d8217000d9f510c75a9c4ac25dd08ed01b612136d534442ece500e55d7f463dd" gracePeriod=30 Mar 21 08:48:58 crc kubenswrapper[4696]: I0321 08:48:58.950457 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="02fca1e0-3685-4d5b-86f9-bef0511876df" containerName="glance-httpd" containerID="cri-o://da6a0ddb2bd8efcc6a441f66dfa4a515e7e3c637ea78a4a987348b985d86dd9d" gracePeriod=30 Mar 21 08:48:59 crc kubenswrapper[4696]: I0321 08:48:59.871063 4696 generic.go:334] "Generic (PLEG): container finished" podID="02fca1e0-3685-4d5b-86f9-bef0511876df" containerID="d8217000d9f510c75a9c4ac25dd08ed01b612136d534442ece500e55d7f463dd" exitCode=143 Mar 21 08:48:59 crc kubenswrapper[4696]: I0321 08:48:59.871116 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"02fca1e0-3685-4d5b-86f9-bef0511876df","Type":"ContainerDied","Data":"d8217000d9f510c75a9c4ac25dd08ed01b612136d534442ece500e55d7f463dd"} Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.380693 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.511283 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-scripts\") pod \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.511343 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f273604b-ebe8-4818-8d9c-322bd8dcfe93-run-httpd\") pod \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.511505 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-combined-ca-bundle\") pod \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.511589 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f273604b-ebe8-4818-8d9c-322bd8dcfe93-log-httpd\") pod \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.511683 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhg6k\" (UniqueName: \"kubernetes.io/projected/f273604b-ebe8-4818-8d9c-322bd8dcfe93-kube-api-access-lhg6k\") pod \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.511734 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-config-data\") pod \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.511763 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-sg-core-conf-yaml\") pod \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\" (UID: \"f273604b-ebe8-4818-8d9c-322bd8dcfe93\") " Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.511992 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f273604b-ebe8-4818-8d9c-322bd8dcfe93-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f273604b-ebe8-4818-8d9c-322bd8dcfe93" (UID: "f273604b-ebe8-4818-8d9c-322bd8dcfe93"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.512765 4696 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f273604b-ebe8-4818-8d9c-322bd8dcfe93-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.513986 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f273604b-ebe8-4818-8d9c-322bd8dcfe93-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f273604b-ebe8-4818-8d9c-322bd8dcfe93" (UID: "f273604b-ebe8-4818-8d9c-322bd8dcfe93"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.521119 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f273604b-ebe8-4818-8d9c-322bd8dcfe93-kube-api-access-lhg6k" (OuterVolumeSpecName: "kube-api-access-lhg6k") pod "f273604b-ebe8-4818-8d9c-322bd8dcfe93" (UID: "f273604b-ebe8-4818-8d9c-322bd8dcfe93"). InnerVolumeSpecName "kube-api-access-lhg6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.527048 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-scripts" (OuterVolumeSpecName: "scripts") pod "f273604b-ebe8-4818-8d9c-322bd8dcfe93" (UID: "f273604b-ebe8-4818-8d9c-322bd8dcfe93"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.568963 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f273604b-ebe8-4818-8d9c-322bd8dcfe93" (UID: "f273604b-ebe8-4818-8d9c-322bd8dcfe93"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.614658 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.614694 4696 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f273604b-ebe8-4818-8d9c-322bd8dcfe93-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.614703 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhg6k\" (UniqueName: \"kubernetes.io/projected/f273604b-ebe8-4818-8d9c-322bd8dcfe93-kube-api-access-lhg6k\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.614714 4696 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.621903 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f273604b-ebe8-4818-8d9c-322bd8dcfe93" (UID: "f273604b-ebe8-4818-8d9c-322bd8dcfe93"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.635954 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-config-data" (OuterVolumeSpecName: "config-data") pod "f273604b-ebe8-4818-8d9c-322bd8dcfe93" (UID: "f273604b-ebe8-4818-8d9c-322bd8dcfe93"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.717579 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.717630 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f273604b-ebe8-4818-8d9c-322bd8dcfe93-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.903562 4696 generic.go:334] "Generic (PLEG): container finished" podID="f273604b-ebe8-4818-8d9c-322bd8dcfe93" containerID="302d82ab7099d32e78e67260e0cc48405091bcaaa141d5555faaa5e07d801479" exitCode=0 Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.903615 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f273604b-ebe8-4818-8d9c-322bd8dcfe93","Type":"ContainerDied","Data":"302d82ab7099d32e78e67260e0cc48405091bcaaa141d5555faaa5e07d801479"} Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.903651 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f273604b-ebe8-4818-8d9c-322bd8dcfe93","Type":"ContainerDied","Data":"36a2a47ecfd009b853f3d409174c86938f21d43d270c800ee42c4255f17898ec"} Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.903671 4696 scope.go:117] "RemoveContainer" containerID="5839a6fcfbe10b15e983b0c9e75e9258e17020dc98ee4a1dc8cc88f9c2f4d14d" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.903843 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.932400 4696 scope.go:117] "RemoveContainer" containerID="8c44d93273437f52d1efdbc780e083274abd6cc990ac7c6831b2bdd5ece0fc89" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.951898 4696 scope.go:117] "RemoveContainer" containerID="139de8563bb4dbb9b04883c62b1ac4af298b15486793bf129d01e835f9cc2a61" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.968600 4696 scope.go:117] "RemoveContainer" containerID="302d82ab7099d32e78e67260e0cc48405091bcaaa141d5555faaa5e07d801479" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.984850 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.996532 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.997955 4696 scope.go:117] "RemoveContainer" containerID="5839a6fcfbe10b15e983b0c9e75e9258e17020dc98ee4a1dc8cc88f9c2f4d14d" Mar 21 08:49:00 crc kubenswrapper[4696]: E0321 08:49:00.998400 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5839a6fcfbe10b15e983b0c9e75e9258e17020dc98ee4a1dc8cc88f9c2f4d14d\": container with ID starting with 5839a6fcfbe10b15e983b0c9e75e9258e17020dc98ee4a1dc8cc88f9c2f4d14d not found: ID does not exist" containerID="5839a6fcfbe10b15e983b0c9e75e9258e17020dc98ee4a1dc8cc88f9c2f4d14d" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.998486 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5839a6fcfbe10b15e983b0c9e75e9258e17020dc98ee4a1dc8cc88f9c2f4d14d"} err="failed to get container status \"5839a6fcfbe10b15e983b0c9e75e9258e17020dc98ee4a1dc8cc88f9c2f4d14d\": rpc error: code = NotFound desc = could not find container \"5839a6fcfbe10b15e983b0c9e75e9258e17020dc98ee4a1dc8cc88f9c2f4d14d\": container with ID starting with 5839a6fcfbe10b15e983b0c9e75e9258e17020dc98ee4a1dc8cc88f9c2f4d14d not found: ID does not exist" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.998559 4696 scope.go:117] "RemoveContainer" containerID="8c44d93273437f52d1efdbc780e083274abd6cc990ac7c6831b2bdd5ece0fc89" Mar 21 08:49:00 crc kubenswrapper[4696]: E0321 08:49:00.999141 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c44d93273437f52d1efdbc780e083274abd6cc990ac7c6831b2bdd5ece0fc89\": container with ID starting with 8c44d93273437f52d1efdbc780e083274abd6cc990ac7c6831b2bdd5ece0fc89 not found: ID does not exist" containerID="8c44d93273437f52d1efdbc780e083274abd6cc990ac7c6831b2bdd5ece0fc89" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.999196 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c44d93273437f52d1efdbc780e083274abd6cc990ac7c6831b2bdd5ece0fc89"} err="failed to get container status \"8c44d93273437f52d1efdbc780e083274abd6cc990ac7c6831b2bdd5ece0fc89\": rpc error: code = NotFound desc = could not find container \"8c44d93273437f52d1efdbc780e083274abd6cc990ac7c6831b2bdd5ece0fc89\": container with ID starting with 8c44d93273437f52d1efdbc780e083274abd6cc990ac7c6831b2bdd5ece0fc89 not found: ID does not exist" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.999231 4696 scope.go:117] "RemoveContainer" containerID="139de8563bb4dbb9b04883c62b1ac4af298b15486793bf129d01e835f9cc2a61" Mar 21 08:49:00 crc kubenswrapper[4696]: E0321 08:49:00.999782 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"139de8563bb4dbb9b04883c62b1ac4af298b15486793bf129d01e835f9cc2a61\": container with ID starting with 139de8563bb4dbb9b04883c62b1ac4af298b15486793bf129d01e835f9cc2a61 not found: ID does not exist" containerID="139de8563bb4dbb9b04883c62b1ac4af298b15486793bf129d01e835f9cc2a61" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.999844 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"139de8563bb4dbb9b04883c62b1ac4af298b15486793bf129d01e835f9cc2a61"} err="failed to get container status \"139de8563bb4dbb9b04883c62b1ac4af298b15486793bf129d01e835f9cc2a61\": rpc error: code = NotFound desc = could not find container \"139de8563bb4dbb9b04883c62b1ac4af298b15486793bf129d01e835f9cc2a61\": container with ID starting with 139de8563bb4dbb9b04883c62b1ac4af298b15486793bf129d01e835f9cc2a61 not found: ID does not exist" Mar 21 08:49:00 crc kubenswrapper[4696]: I0321 08:49:00.999884 4696 scope.go:117] "RemoveContainer" containerID="302d82ab7099d32e78e67260e0cc48405091bcaaa141d5555faaa5e07d801479" Mar 21 08:49:01 crc kubenswrapper[4696]: E0321 08:49:01.013681 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"302d82ab7099d32e78e67260e0cc48405091bcaaa141d5555faaa5e07d801479\": container with ID starting with 302d82ab7099d32e78e67260e0cc48405091bcaaa141d5555faaa5e07d801479 not found: ID does not exist" containerID="302d82ab7099d32e78e67260e0cc48405091bcaaa141d5555faaa5e07d801479" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.013721 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"302d82ab7099d32e78e67260e0cc48405091bcaaa141d5555faaa5e07d801479"} err="failed to get container status \"302d82ab7099d32e78e67260e0cc48405091bcaaa141d5555faaa5e07d801479\": rpc error: code = NotFound desc = could not find container \"302d82ab7099d32e78e67260e0cc48405091bcaaa141d5555faaa5e07d801479\": container with ID starting with 302d82ab7099d32e78e67260e0cc48405091bcaaa141d5555faaa5e07d801479 not found: ID does not exist" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.014741 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:49:01 crc kubenswrapper[4696]: E0321 08:49:01.015247 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f273604b-ebe8-4818-8d9c-322bd8dcfe93" containerName="ceilometer-central-agent" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.015275 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f273604b-ebe8-4818-8d9c-322bd8dcfe93" containerName="ceilometer-central-agent" Mar 21 08:49:01 crc kubenswrapper[4696]: E0321 08:49:01.015295 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f273604b-ebe8-4818-8d9c-322bd8dcfe93" containerName="ceilometer-notification-agent" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.015305 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f273604b-ebe8-4818-8d9c-322bd8dcfe93" containerName="ceilometer-notification-agent" Mar 21 08:49:01 crc kubenswrapper[4696]: E0321 08:49:01.015326 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f273604b-ebe8-4818-8d9c-322bd8dcfe93" containerName="sg-core" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.015334 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f273604b-ebe8-4818-8d9c-322bd8dcfe93" containerName="sg-core" Mar 21 08:49:01 crc kubenswrapper[4696]: E0321 08:49:01.015348 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f273604b-ebe8-4818-8d9c-322bd8dcfe93" containerName="proxy-httpd" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.015356 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f273604b-ebe8-4818-8d9c-322bd8dcfe93" containerName="proxy-httpd" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.015577 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f273604b-ebe8-4818-8d9c-322bd8dcfe93" containerName="proxy-httpd" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.015599 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f273604b-ebe8-4818-8d9c-322bd8dcfe93" containerName="ceilometer-central-agent" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.015635 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f273604b-ebe8-4818-8d9c-322bd8dcfe93" containerName="sg-core" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.015649 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f273604b-ebe8-4818-8d9c-322bd8dcfe93" containerName="ceilometer-notification-agent" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.019339 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.021510 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.023421 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.028324 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.126388 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.126757 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.126887 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee154bce-bf2e-4b07-81d9-f2178f71973a-run-httpd\") pod \"ceilometer-0\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.126907 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee154bce-bf2e-4b07-81d9-f2178f71973a-log-httpd\") pod \"ceilometer-0\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.126929 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47lhx\" (UniqueName: \"kubernetes.io/projected/ee154bce-bf2e-4b07-81d9-f2178f71973a-kube-api-access-47lhx\") pod \"ceilometer-0\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.126958 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-config-data\") pod \"ceilometer-0\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.126973 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-scripts\") pod \"ceilometer-0\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.229264 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.229391 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee154bce-bf2e-4b07-81d9-f2178f71973a-run-httpd\") pod \"ceilometer-0\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.229419 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee154bce-bf2e-4b07-81d9-f2178f71973a-log-httpd\") pod \"ceilometer-0\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.229446 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47lhx\" (UniqueName: \"kubernetes.io/projected/ee154bce-bf2e-4b07-81d9-f2178f71973a-kube-api-access-47lhx\") pod \"ceilometer-0\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.229502 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-config-data\") pod \"ceilometer-0\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.229522 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-scripts\") pod \"ceilometer-0\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.229673 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.229847 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee154bce-bf2e-4b07-81d9-f2178f71973a-run-httpd\") pod \"ceilometer-0\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.230281 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee154bce-bf2e-4b07-81d9-f2178f71973a-log-httpd\") pod \"ceilometer-0\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.236517 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.236795 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-scripts\") pod \"ceilometer-0\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.236803 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.237168 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-config-data\") pod \"ceilometer-0\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.249697 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47lhx\" (UniqueName: \"kubernetes.io/projected/ee154bce-bf2e-4b07-81d9-f2178f71973a-kube-api-access-47lhx\") pod \"ceilometer-0\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.338206 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.703543 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.838241 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.917104 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee154bce-bf2e-4b07-81d9-f2178f71973a","Type":"ContainerStarted","Data":"f88c15c41b0306cb7a58f39f784e32361c971beb5b5af33efce45ed4f1a3a785"} Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.920985 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1af36c17-5501-4d4e-87a9-e9b43cb9eef0" containerName="glance-log" containerID="cri-o://c3f089a13d2570c4fe4cafb5efecab2fdcd188b1e190e48ec3fb875343fe9ab2" gracePeriod=30 Mar 21 08:49:01 crc kubenswrapper[4696]: I0321 08:49:01.921541 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1af36c17-5501-4d4e-87a9-e9b43cb9eef0" containerName="glance-httpd" containerID="cri-o://530a9760edef3bd29be8445624f8167d7d983c70e52b38653480e80e7a5c7b3a" gracePeriod=30 Mar 21 08:49:02 crc kubenswrapper[4696]: I0321 08:49:02.545658 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f273604b-ebe8-4818-8d9c-322bd8dcfe93" path="/var/lib/kubelet/pods/f273604b-ebe8-4818-8d9c-322bd8dcfe93/volumes" Mar 21 08:49:02 crc kubenswrapper[4696]: I0321 08:49:02.715336 4696 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","poda32ac807-7955-425a-a971-e07ddde36bb0"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort poda32ac807-7955-425a-a971-e07ddde36bb0] : Timed out while waiting for systemd to remove kubepods-besteffort-poda32ac807_7955_425a_a971_e07ddde36bb0.slice" Mar 21 08:49:02 crc kubenswrapper[4696]: I0321 08:49:02.932514 4696 generic.go:334] "Generic (PLEG): container finished" podID="02fca1e0-3685-4d5b-86f9-bef0511876df" containerID="da6a0ddb2bd8efcc6a441f66dfa4a515e7e3c637ea78a4a987348b985d86dd9d" exitCode=0 Mar 21 08:49:02 crc kubenswrapper[4696]: I0321 08:49:02.932583 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"02fca1e0-3685-4d5b-86f9-bef0511876df","Type":"ContainerDied","Data":"da6a0ddb2bd8efcc6a441f66dfa4a515e7e3c637ea78a4a987348b985d86dd9d"} Mar 21 08:49:02 crc kubenswrapper[4696]: I0321 08:49:02.934386 4696 generic.go:334] "Generic (PLEG): container finished" podID="1af36c17-5501-4d4e-87a9-e9b43cb9eef0" containerID="c3f089a13d2570c4fe4cafb5efecab2fdcd188b1e190e48ec3fb875343fe9ab2" exitCode=143 Mar 21 08:49:02 crc kubenswrapper[4696]: I0321 08:49:02.934424 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1af36c17-5501-4d4e-87a9-e9b43cb9eef0","Type":"ContainerDied","Data":"c3f089a13d2570c4fe4cafb5efecab2fdcd188b1e190e48ec3fb875343fe9ab2"} Mar 21 08:49:02 crc kubenswrapper[4696]: I0321 08:49:02.936585 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee154bce-bf2e-4b07-81d9-f2178f71973a","Type":"ContainerStarted","Data":"56edcf19c7b5e1e663a3e748a2ce752c2f1fcccf530bf4e57a55cdc8db2ffe27"} Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.172553 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.279076 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xpbn\" (UniqueName: \"kubernetes.io/projected/02fca1e0-3685-4d5b-86f9-bef0511876df-kube-api-access-6xpbn\") pod \"02fca1e0-3685-4d5b-86f9-bef0511876df\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.279437 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-scripts\") pod \"02fca1e0-3685-4d5b-86f9-bef0511876df\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.279483 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/02fca1e0-3685-4d5b-86f9-bef0511876df-httpd-run\") pod \"02fca1e0-3685-4d5b-86f9-bef0511876df\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.279529 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02fca1e0-3685-4d5b-86f9-bef0511876df-logs\") pod \"02fca1e0-3685-4d5b-86f9-bef0511876df\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.279545 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-config-data\") pod \"02fca1e0-3685-4d5b-86f9-bef0511876df\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.279630 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-public-tls-certs\") pod \"02fca1e0-3685-4d5b-86f9-bef0511876df\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.279650 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-combined-ca-bundle\") pod \"02fca1e0-3685-4d5b-86f9-bef0511876df\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.280292 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02fca1e0-3685-4d5b-86f9-bef0511876df-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "02fca1e0-3685-4d5b-86f9-bef0511876df" (UID: "02fca1e0-3685-4d5b-86f9-bef0511876df"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.280583 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") pod \"02fca1e0-3685-4d5b-86f9-bef0511876df\" (UID: \"02fca1e0-3685-4d5b-86f9-bef0511876df\") " Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.281485 4696 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/02fca1e0-3685-4d5b-86f9-bef0511876df-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.281895 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02fca1e0-3685-4d5b-86f9-bef0511876df-logs" (OuterVolumeSpecName: "logs") pod "02fca1e0-3685-4d5b-86f9-bef0511876df" (UID: "02fca1e0-3685-4d5b-86f9-bef0511876df"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.296264 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-scripts" (OuterVolumeSpecName: "scripts") pod "02fca1e0-3685-4d5b-86f9-bef0511876df" (UID: "02fca1e0-3685-4d5b-86f9-bef0511876df"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.297885 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02fca1e0-3685-4d5b-86f9-bef0511876df-kube-api-access-6xpbn" (OuterVolumeSpecName: "kube-api-access-6xpbn") pod "02fca1e0-3685-4d5b-86f9-bef0511876df" (UID: "02fca1e0-3685-4d5b-86f9-bef0511876df"). InnerVolumeSpecName "kube-api-access-6xpbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.316203 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689" (OuterVolumeSpecName: "glance") pod "02fca1e0-3685-4d5b-86f9-bef0511876df" (UID: "02fca1e0-3685-4d5b-86f9-bef0511876df"). InnerVolumeSpecName "pvc-984a8dcb-40a1-4f39-bf33-0005b91df689". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.318425 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02fca1e0-3685-4d5b-86f9-bef0511876df" (UID: "02fca1e0-3685-4d5b-86f9-bef0511876df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.372950 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-config-data" (OuterVolumeSpecName: "config-data") pod "02fca1e0-3685-4d5b-86f9-bef0511876df" (UID: "02fca1e0-3685-4d5b-86f9-bef0511876df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.374122 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "02fca1e0-3685-4d5b-86f9-bef0511876df" (UID: "02fca1e0-3685-4d5b-86f9-bef0511876df"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.383348 4696 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.383378 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.383404 4696 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") on node \"crc\" " Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.383416 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xpbn\" (UniqueName: \"kubernetes.io/projected/02fca1e0-3685-4d5b-86f9-bef0511876df-kube-api-access-6xpbn\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.383425 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.383434 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02fca1e0-3685-4d5b-86f9-bef0511876df-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.383446 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02fca1e0-3685-4d5b-86f9-bef0511876df-logs\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.420166 4696 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.420311 4696 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-984a8dcb-40a1-4f39-bf33-0005b91df689" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689") on node "crc" Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.485038 4696 reconciler_common.go:293] "Volume detached for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.711096 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.960253 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee154bce-bf2e-4b07-81d9-f2178f71973a","Type":"ContainerStarted","Data":"193a3d19134bb486e9301c5545a00e7461f8c3beb00ed97011d47dd42540b7d2"} Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.966474 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"02fca1e0-3685-4d5b-86f9-bef0511876df","Type":"ContainerDied","Data":"e660b83e06e17513aae66d0ffece19e7b8822ee6ee90e63e0664bbe24152e3b1"} Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.966510 4696 scope.go:117] "RemoveContainer" containerID="da6a0ddb2bd8efcc6a441f66dfa4a515e7e3c637ea78a4a987348b985d86dd9d" Mar 21 08:49:03 crc kubenswrapper[4696]: I0321 08:49:03.966664 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.098774 4696 scope.go:117] "RemoveContainer" containerID="d8217000d9f510c75a9c4ac25dd08ed01b612136d534442ece500e55d7f463dd" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.132741 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.162165 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.180567 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 08:49:04 crc kubenswrapper[4696]: E0321 08:49:04.181055 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02fca1e0-3685-4d5b-86f9-bef0511876df" containerName="glance-httpd" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.181073 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="02fca1e0-3685-4d5b-86f9-bef0511876df" containerName="glance-httpd" Mar 21 08:49:04 crc kubenswrapper[4696]: E0321 08:49:04.181098 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02fca1e0-3685-4d5b-86f9-bef0511876df" containerName="glance-log" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.181104 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="02fca1e0-3685-4d5b-86f9-bef0511876df" containerName="glance-log" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.181319 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="02fca1e0-3685-4d5b-86f9-bef0511876df" containerName="glance-httpd" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.181350 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="02fca1e0-3685-4d5b-86f9-bef0511876df" containerName="glance-log" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.182450 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.189619 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.189857 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.204208 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.300178 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/325e7183-8d8e-4a92-ab49-3dc4a043c21c-logs\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.300262 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.300349 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/325e7183-8d8e-4a92-ab49-3dc4a043c21c-scripts\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.300466 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/325e7183-8d8e-4a92-ab49-3dc4a043c21c-config-data\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.300600 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kb62s\" (UniqueName: \"kubernetes.io/projected/325e7183-8d8e-4a92-ab49-3dc4a043c21c-kube-api-access-kb62s\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.300670 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/325e7183-8d8e-4a92-ab49-3dc4a043c21c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.300778 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/325e7183-8d8e-4a92-ab49-3dc4a043c21c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.300805 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/325e7183-8d8e-4a92-ab49-3dc4a043c21c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.402313 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kb62s\" (UniqueName: \"kubernetes.io/projected/325e7183-8d8e-4a92-ab49-3dc4a043c21c-kube-api-access-kb62s\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.402388 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/325e7183-8d8e-4a92-ab49-3dc4a043c21c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.402464 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/325e7183-8d8e-4a92-ab49-3dc4a043c21c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.402487 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/325e7183-8d8e-4a92-ab49-3dc4a043c21c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.403067 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/325e7183-8d8e-4a92-ab49-3dc4a043c21c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.403141 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/325e7183-8d8e-4a92-ab49-3dc4a043c21c-logs\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.403361 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/325e7183-8d8e-4a92-ab49-3dc4a043c21c-logs\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.403434 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.403462 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/325e7183-8d8e-4a92-ab49-3dc4a043c21c-scripts\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.403534 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/325e7183-8d8e-4a92-ab49-3dc4a043c21c-config-data\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.407520 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/325e7183-8d8e-4a92-ab49-3dc4a043c21c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.408137 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/325e7183-8d8e-4a92-ab49-3dc4a043c21c-scripts\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.410234 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.410261 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0bb60238debc80c6ed39528f98bb598b47b2c7acc3e22f58b887f712b724eea4/globalmount\"" pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.410788 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/325e7183-8d8e-4a92-ab49-3dc4a043c21c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.411312 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/325e7183-8d8e-4a92-ab49-3dc4a043c21c-config-data\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.423501 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kb62s\" (UniqueName: \"kubernetes.io/projected/325e7183-8d8e-4a92-ab49-3dc4a043c21c-kube-api-access-kb62s\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.462507 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-984a8dcb-40a1-4f39-bf33-0005b91df689\") pod \"glance-default-external-api-0\" (UID: \"325e7183-8d8e-4a92-ab49-3dc4a043c21c\") " pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.520783 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.550207 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02fca1e0-3685-4d5b-86f9-bef0511876df" path="/var/lib/kubelet/pods/02fca1e0-3685-4d5b-86f9-bef0511876df/volumes" Mar 21 08:49:04 crc kubenswrapper[4696]: I0321 08:49:04.985260 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee154bce-bf2e-4b07-81d9-f2178f71973a","Type":"ContainerStarted","Data":"09dc39ceb111d4ebcced1a1a7c1b527465533c6bc4ad8f2d2086bdc85a5b3bf4"} Mar 21 08:49:05 crc kubenswrapper[4696]: I0321 08:49:05.290878 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 21 08:49:05 crc kubenswrapper[4696]: I0321 08:49:05.989264 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.005564 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"325e7183-8d8e-4a92-ab49-3dc4a043c21c","Type":"ContainerStarted","Data":"e518a06d197a49a6fd57a5a9215db93c05f9af96bff566e60a27bf3429ce95cf"} Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.031141 4696 generic.go:334] "Generic (PLEG): container finished" podID="1af36c17-5501-4d4e-87a9-e9b43cb9eef0" containerID="530a9760edef3bd29be8445624f8167d7d983c70e52b38653480e80e7a5c7b3a" exitCode=0 Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.031433 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1af36c17-5501-4d4e-87a9-e9b43cb9eef0","Type":"ContainerDied","Data":"530a9760edef3bd29be8445624f8167d7d983c70e52b38653480e80e7a5c7b3a"} Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.031517 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1af36c17-5501-4d4e-87a9-e9b43cb9eef0","Type":"ContainerDied","Data":"3a9bab0c350025ca189e82a6631725e712e1e13767d1bddd512764b878f9a05e"} Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.031586 4696 scope.go:117] "RemoveContainer" containerID="530a9760edef3bd29be8445624f8167d7d983c70e52b38653480e80e7a5c7b3a" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.031862 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.069879 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") pod \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.069928 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-combined-ca-bundle\") pod \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.070037 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-httpd-run\") pod \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.070112 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-internal-tls-certs\") pod \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.070138 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-config-data\") pod \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.070174 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6n56\" (UniqueName: \"kubernetes.io/projected/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-kube-api-access-v6n56\") pod \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.070247 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-scripts\") pod \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.070341 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-logs\") pod \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\" (UID: \"1af36c17-5501-4d4e-87a9-e9b43cb9eef0\") " Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.070881 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1af36c17-5501-4d4e-87a9-e9b43cb9eef0" (UID: "1af36c17-5501-4d4e-87a9-e9b43cb9eef0"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.071254 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-logs" (OuterVolumeSpecName: "logs") pod "1af36c17-5501-4d4e-87a9-e9b43cb9eef0" (UID: "1af36c17-5501-4d4e-87a9-e9b43cb9eef0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.075073 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-kube-api-access-v6n56" (OuterVolumeSpecName: "kube-api-access-v6n56") pod "1af36c17-5501-4d4e-87a9-e9b43cb9eef0" (UID: "1af36c17-5501-4d4e-87a9-e9b43cb9eef0"). InnerVolumeSpecName "kube-api-access-v6n56". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.080798 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-scripts" (OuterVolumeSpecName: "scripts") pod "1af36c17-5501-4d4e-87a9-e9b43cb9eef0" (UID: "1af36c17-5501-4d4e-87a9-e9b43cb9eef0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.081199 4696 scope.go:117] "RemoveContainer" containerID="c3f089a13d2570c4fe4cafb5efecab2fdcd188b1e190e48ec3fb875343fe9ab2" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.118744 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205" (OuterVolumeSpecName: "glance") pod "1af36c17-5501-4d4e-87a9-e9b43cb9eef0" (UID: "1af36c17-5501-4d4e-87a9-e9b43cb9eef0"). InnerVolumeSpecName "pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.125952 4696 scope.go:117] "RemoveContainer" containerID="530a9760edef3bd29be8445624f8167d7d983c70e52b38653480e80e7a5c7b3a" Mar 21 08:49:06 crc kubenswrapper[4696]: E0321 08:49:06.129932 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"530a9760edef3bd29be8445624f8167d7d983c70e52b38653480e80e7a5c7b3a\": container with ID starting with 530a9760edef3bd29be8445624f8167d7d983c70e52b38653480e80e7a5c7b3a not found: ID does not exist" containerID="530a9760edef3bd29be8445624f8167d7d983c70e52b38653480e80e7a5c7b3a" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.129975 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"530a9760edef3bd29be8445624f8167d7d983c70e52b38653480e80e7a5c7b3a"} err="failed to get container status \"530a9760edef3bd29be8445624f8167d7d983c70e52b38653480e80e7a5c7b3a\": rpc error: code = NotFound desc = could not find container \"530a9760edef3bd29be8445624f8167d7d983c70e52b38653480e80e7a5c7b3a\": container with ID starting with 530a9760edef3bd29be8445624f8167d7d983c70e52b38653480e80e7a5c7b3a not found: ID does not exist" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.130000 4696 scope.go:117] "RemoveContainer" containerID="c3f089a13d2570c4fe4cafb5efecab2fdcd188b1e190e48ec3fb875343fe9ab2" Mar 21 08:49:06 crc kubenswrapper[4696]: E0321 08:49:06.130389 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3f089a13d2570c4fe4cafb5efecab2fdcd188b1e190e48ec3fb875343fe9ab2\": container with ID starting with c3f089a13d2570c4fe4cafb5efecab2fdcd188b1e190e48ec3fb875343fe9ab2 not found: ID does not exist" containerID="c3f089a13d2570c4fe4cafb5efecab2fdcd188b1e190e48ec3fb875343fe9ab2" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.130409 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3f089a13d2570c4fe4cafb5efecab2fdcd188b1e190e48ec3fb875343fe9ab2"} err="failed to get container status \"c3f089a13d2570c4fe4cafb5efecab2fdcd188b1e190e48ec3fb875343fe9ab2\": rpc error: code = NotFound desc = could not find container \"c3f089a13d2570c4fe4cafb5efecab2fdcd188b1e190e48ec3fb875343fe9ab2\": container with ID starting with c3f089a13d2570c4fe4cafb5efecab2fdcd188b1e190e48ec3fb875343fe9ab2 not found: ID does not exist" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.131716 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1af36c17-5501-4d4e-87a9-e9b43cb9eef0" (UID: "1af36c17-5501-4d4e-87a9-e9b43cb9eef0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.161935 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1af36c17-5501-4d4e-87a9-e9b43cb9eef0" (UID: "1af36c17-5501-4d4e-87a9-e9b43cb9eef0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.176242 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6n56\" (UniqueName: \"kubernetes.io/projected/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-kube-api-access-v6n56\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.177027 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.177135 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-logs\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.177209 4696 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") on node \"crc\" " Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.177268 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.177332 4696 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.177388 4696 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.209371 4696 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.209501 4696 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205") on node "crc" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.226576 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-config-data" (OuterVolumeSpecName: "config-data") pod "1af36c17-5501-4d4e-87a9-e9b43cb9eef0" (UID: "1af36c17-5501-4d4e-87a9-e9b43cb9eef0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.278774 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1af36c17-5501-4d4e-87a9-e9b43cb9eef0-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.278805 4696 reconciler_common.go:293] "Volume detached for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.380748 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.395705 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.408330 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 08:49:06 crc kubenswrapper[4696]: E0321 08:49:06.409072 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1af36c17-5501-4d4e-87a9-e9b43cb9eef0" containerName="glance-httpd" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.409092 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1af36c17-5501-4d4e-87a9-e9b43cb9eef0" containerName="glance-httpd" Mar 21 08:49:06 crc kubenswrapper[4696]: E0321 08:49:06.409116 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1af36c17-5501-4d4e-87a9-e9b43cb9eef0" containerName="glance-log" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.409125 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1af36c17-5501-4d4e-87a9-e9b43cb9eef0" containerName="glance-log" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.409376 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1af36c17-5501-4d4e-87a9-e9b43cb9eef0" containerName="glance-httpd" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.409402 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1af36c17-5501-4d4e-87a9-e9b43cb9eef0" containerName="glance-log" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.411592 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.420129 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.420622 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.423357 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.493262 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.493347 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wl4m\" (UniqueName: \"kubernetes.io/projected/a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7-kube-api-access-8wl4m\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.493398 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.493438 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.493466 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.493494 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.493561 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7-logs\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.493655 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.560261 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1af36c17-5501-4d4e-87a9-e9b43cb9eef0" path="/var/lib/kubelet/pods/1af36c17-5501-4d4e-87a9-e9b43cb9eef0/volumes" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.594825 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.594885 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.594929 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wl4m\" (UniqueName: \"kubernetes.io/projected/a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7-kube-api-access-8wl4m\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.594963 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.594991 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.595013 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.595032 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.595079 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7-logs\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.595514 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7-logs\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.604343 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.632472 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.632784 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.633443 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.634511 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.641526 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wl4m\" (UniqueName: \"kubernetes.io/projected/a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7-kube-api-access-8wl4m\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.654779 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.654860 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/36feb21b6af5aad5e90205d24e3f1d239a0e9aaaadf2ca8cf7f1809b63388e1a/globalmount\"" pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.714204 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68056206-8a8e-4de9-8ab6-e52e14c2e205\") pod \"glance-default-internal-api-0\" (UID: \"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7\") " pod="openstack/glance-default-internal-api-0" Mar 21 08:49:06 crc kubenswrapper[4696]: I0321 08:49:06.824268 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 21 08:49:07 crc kubenswrapper[4696]: I0321 08:49:07.053243 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"325e7183-8d8e-4a92-ab49-3dc4a043c21c","Type":"ContainerStarted","Data":"491497c640381c562baf83426109595278c4202fb71c6b6e4d1ab849aea43b3e"} Mar 21 08:49:07 crc kubenswrapper[4696]: I0321 08:49:07.075072 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee154bce-bf2e-4b07-81d9-f2178f71973a","Type":"ContainerStarted","Data":"11b947598c7c8c0c8df9873f444b0ab7bbaa43d2d60fc6a236874da075946586"} Mar 21 08:49:07 crc kubenswrapper[4696]: I0321 08:49:07.075269 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee154bce-bf2e-4b07-81d9-f2178f71973a" containerName="ceilometer-central-agent" containerID="cri-o://56edcf19c7b5e1e663a3e748a2ce752c2f1fcccf530bf4e57a55cdc8db2ffe27" gracePeriod=30 Mar 21 08:49:07 crc kubenswrapper[4696]: I0321 08:49:07.075554 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 21 08:49:07 crc kubenswrapper[4696]: I0321 08:49:07.075852 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee154bce-bf2e-4b07-81d9-f2178f71973a" containerName="proxy-httpd" containerID="cri-o://11b947598c7c8c0c8df9873f444b0ab7bbaa43d2d60fc6a236874da075946586" gracePeriod=30 Mar 21 08:49:07 crc kubenswrapper[4696]: I0321 08:49:07.075895 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee154bce-bf2e-4b07-81d9-f2178f71973a" containerName="sg-core" containerID="cri-o://09dc39ceb111d4ebcced1a1a7c1b527465533c6bc4ad8f2d2086bdc85a5b3bf4" gracePeriod=30 Mar 21 08:49:07 crc kubenswrapper[4696]: I0321 08:49:07.075928 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee154bce-bf2e-4b07-81d9-f2178f71973a" containerName="ceilometer-notification-agent" containerID="cri-o://193a3d19134bb486e9301c5545a00e7461f8c3beb00ed97011d47dd42540b7d2" gracePeriod=30 Mar 21 08:49:07 crc kubenswrapper[4696]: I0321 08:49:07.467078 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.528540085 podStartE2EDuration="7.467062418s" podCreationTimestamp="2026-03-21 08:49:00 +0000 UTC" firstStartedPulling="2026-03-21 08:49:01.838051472 +0000 UTC m=+1275.958932195" lastFinishedPulling="2026-03-21 08:49:05.776573815 +0000 UTC m=+1279.897454528" observedRunningTime="2026-03-21 08:49:07.10599804 +0000 UTC m=+1281.226878763" watchObservedRunningTime="2026-03-21 08:49:07.467062418 +0000 UTC m=+1281.587943131" Mar 21 08:49:07 crc kubenswrapper[4696]: I0321 08:49:07.474283 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 21 08:49:08 crc kubenswrapper[4696]: I0321 08:49:08.119136 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"325e7183-8d8e-4a92-ab49-3dc4a043c21c","Type":"ContainerStarted","Data":"ca3e37c0c0adb9b8fc3b128c718923f917e94c4aabbb1b8ba2505143ecb48ce1"} Mar 21 08:49:08 crc kubenswrapper[4696]: I0321 08:49:08.126505 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7","Type":"ContainerStarted","Data":"aa055016f88b55b10149dc3ef70458358be7e80d2b2314ed7fb904ba4ca323e2"} Mar 21 08:49:08 crc kubenswrapper[4696]: I0321 08:49:08.126556 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7","Type":"ContainerStarted","Data":"bf29c19d45045a2c1694491c5bf0d811c67047f014b51c0f98a687fd80294672"} Mar 21 08:49:08 crc kubenswrapper[4696]: I0321 08:49:08.132800 4696 generic.go:334] "Generic (PLEG): container finished" podID="ee154bce-bf2e-4b07-81d9-f2178f71973a" containerID="11b947598c7c8c0c8df9873f444b0ab7bbaa43d2d60fc6a236874da075946586" exitCode=0 Mar 21 08:49:08 crc kubenswrapper[4696]: I0321 08:49:08.132846 4696 generic.go:334] "Generic (PLEG): container finished" podID="ee154bce-bf2e-4b07-81d9-f2178f71973a" containerID="09dc39ceb111d4ebcced1a1a7c1b527465533c6bc4ad8f2d2086bdc85a5b3bf4" exitCode=2 Mar 21 08:49:08 crc kubenswrapper[4696]: I0321 08:49:08.132854 4696 generic.go:334] "Generic (PLEG): container finished" podID="ee154bce-bf2e-4b07-81d9-f2178f71973a" containerID="193a3d19134bb486e9301c5545a00e7461f8c3beb00ed97011d47dd42540b7d2" exitCode=0 Mar 21 08:49:08 crc kubenswrapper[4696]: I0321 08:49:08.132876 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee154bce-bf2e-4b07-81d9-f2178f71973a","Type":"ContainerDied","Data":"11b947598c7c8c0c8df9873f444b0ab7bbaa43d2d60fc6a236874da075946586"} Mar 21 08:49:08 crc kubenswrapper[4696]: I0321 08:49:08.132899 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee154bce-bf2e-4b07-81d9-f2178f71973a","Type":"ContainerDied","Data":"09dc39ceb111d4ebcced1a1a7c1b527465533c6bc4ad8f2d2086bdc85a5b3bf4"} Mar 21 08:49:08 crc kubenswrapper[4696]: I0321 08:49:08.132909 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee154bce-bf2e-4b07-81d9-f2178f71973a","Type":"ContainerDied","Data":"193a3d19134bb486e9301c5545a00e7461f8c3beb00ed97011d47dd42540b7d2"} Mar 21 08:49:08 crc kubenswrapper[4696]: I0321 08:49:08.162256 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.162238058 podStartE2EDuration="4.162238058s" podCreationTimestamp="2026-03-21 08:49:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:49:08.139085588 +0000 UTC m=+1282.259966301" watchObservedRunningTime="2026-03-21 08:49:08.162238058 +0000 UTC m=+1282.283118771" Mar 21 08:49:09 crc kubenswrapper[4696]: I0321 08:49:09.146003 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7","Type":"ContainerStarted","Data":"e77157028fc5ec6066e5beb1386a6e10e8369b9f1e541f573a83acd8469b5877"} Mar 21 08:49:09 crc kubenswrapper[4696]: I0321 08:49:09.171575 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.171559538 podStartE2EDuration="3.171559538s" podCreationTimestamp="2026-03-21 08:49:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:49:09.164608226 +0000 UTC m=+1283.285488929" watchObservedRunningTime="2026-03-21 08:49:09.171559538 +0000 UTC m=+1283.292440241" Mar 21 08:49:11 crc kubenswrapper[4696]: I0321 08:49:11.176570 4696 generic.go:334] "Generic (PLEG): container finished" podID="5e448486-5596-4442-a11a-804a7b7b94d8" containerID="9eae8c5cbd9fc1b2ae5e65811cb3ddc5f1961c88f5db9e6c115709c50c3f3c03" exitCode=0 Mar 21 08:49:11 crc kubenswrapper[4696]: I0321 08:49:11.176911 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-j298g" event={"ID":"5e448486-5596-4442-a11a-804a7b7b94d8","Type":"ContainerDied","Data":"9eae8c5cbd9fc1b2ae5e65811cb3ddc5f1961c88f5db9e6c115709c50c3f3c03"} Mar 21 08:49:12 crc kubenswrapper[4696]: I0321 08:49:12.718792 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-j298g" Mar 21 08:49:12 crc kubenswrapper[4696]: I0321 08:49:12.833211 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e448486-5596-4442-a11a-804a7b7b94d8-scripts\") pod \"5e448486-5596-4442-a11a-804a7b7b94d8\" (UID: \"5e448486-5596-4442-a11a-804a7b7b94d8\") " Mar 21 08:49:12 crc kubenswrapper[4696]: I0321 08:49:12.833287 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m58v5\" (UniqueName: \"kubernetes.io/projected/5e448486-5596-4442-a11a-804a7b7b94d8-kube-api-access-m58v5\") pod \"5e448486-5596-4442-a11a-804a7b7b94d8\" (UID: \"5e448486-5596-4442-a11a-804a7b7b94d8\") " Mar 21 08:49:12 crc kubenswrapper[4696]: I0321 08:49:12.833322 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e448486-5596-4442-a11a-804a7b7b94d8-combined-ca-bundle\") pod \"5e448486-5596-4442-a11a-804a7b7b94d8\" (UID: \"5e448486-5596-4442-a11a-804a7b7b94d8\") " Mar 21 08:49:12 crc kubenswrapper[4696]: I0321 08:49:12.833496 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e448486-5596-4442-a11a-804a7b7b94d8-config-data\") pod \"5e448486-5596-4442-a11a-804a7b7b94d8\" (UID: \"5e448486-5596-4442-a11a-804a7b7b94d8\") " Mar 21 08:49:12 crc kubenswrapper[4696]: I0321 08:49:12.845463 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e448486-5596-4442-a11a-804a7b7b94d8-kube-api-access-m58v5" (OuterVolumeSpecName: "kube-api-access-m58v5") pod "5e448486-5596-4442-a11a-804a7b7b94d8" (UID: "5e448486-5596-4442-a11a-804a7b7b94d8"). InnerVolumeSpecName "kube-api-access-m58v5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:49:12 crc kubenswrapper[4696]: I0321 08:49:12.864074 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e448486-5596-4442-a11a-804a7b7b94d8-scripts" (OuterVolumeSpecName: "scripts") pod "5e448486-5596-4442-a11a-804a7b7b94d8" (UID: "5e448486-5596-4442-a11a-804a7b7b94d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:12 crc kubenswrapper[4696]: I0321 08:49:12.884615 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e448486-5596-4442-a11a-804a7b7b94d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e448486-5596-4442-a11a-804a7b7b94d8" (UID: "5e448486-5596-4442-a11a-804a7b7b94d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:12 crc kubenswrapper[4696]: I0321 08:49:12.901350 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e448486-5596-4442-a11a-804a7b7b94d8-config-data" (OuterVolumeSpecName: "config-data") pod "5e448486-5596-4442-a11a-804a7b7b94d8" (UID: "5e448486-5596-4442-a11a-804a7b7b94d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:12 crc kubenswrapper[4696]: I0321 08:49:12.938290 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e448486-5596-4442-a11a-804a7b7b94d8-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:12 crc kubenswrapper[4696]: I0321 08:49:12.938322 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m58v5\" (UniqueName: \"kubernetes.io/projected/5e448486-5596-4442-a11a-804a7b7b94d8-kube-api-access-m58v5\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:12 crc kubenswrapper[4696]: I0321 08:49:12.938336 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e448486-5596-4442-a11a-804a7b7b94d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:12 crc kubenswrapper[4696]: I0321 08:49:12.938345 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e448486-5596-4442-a11a-804a7b7b94d8-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.031454 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.140610 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee154bce-bf2e-4b07-81d9-f2178f71973a-run-httpd\") pod \"ee154bce-bf2e-4b07-81d9-f2178f71973a\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.140917 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-combined-ca-bundle\") pod \"ee154bce-bf2e-4b07-81d9-f2178f71973a\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.140952 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-sg-core-conf-yaml\") pod \"ee154bce-bf2e-4b07-81d9-f2178f71973a\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.140973 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee154bce-bf2e-4b07-81d9-f2178f71973a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ee154bce-bf2e-4b07-81d9-f2178f71973a" (UID: "ee154bce-bf2e-4b07-81d9-f2178f71973a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.141003 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47lhx\" (UniqueName: \"kubernetes.io/projected/ee154bce-bf2e-4b07-81d9-f2178f71973a-kube-api-access-47lhx\") pod \"ee154bce-bf2e-4b07-81d9-f2178f71973a\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.141103 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee154bce-bf2e-4b07-81d9-f2178f71973a-log-httpd\") pod \"ee154bce-bf2e-4b07-81d9-f2178f71973a\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.141189 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-scripts\") pod \"ee154bce-bf2e-4b07-81d9-f2178f71973a\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.141309 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-config-data\") pod \"ee154bce-bf2e-4b07-81d9-f2178f71973a\" (UID: \"ee154bce-bf2e-4b07-81d9-f2178f71973a\") " Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.141921 4696 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee154bce-bf2e-4b07-81d9-f2178f71973a-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.142749 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee154bce-bf2e-4b07-81d9-f2178f71973a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ee154bce-bf2e-4b07-81d9-f2178f71973a" (UID: "ee154bce-bf2e-4b07-81d9-f2178f71973a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.145125 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-scripts" (OuterVolumeSpecName: "scripts") pod "ee154bce-bf2e-4b07-81d9-f2178f71973a" (UID: "ee154bce-bf2e-4b07-81d9-f2178f71973a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.150115 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee154bce-bf2e-4b07-81d9-f2178f71973a-kube-api-access-47lhx" (OuterVolumeSpecName: "kube-api-access-47lhx") pod "ee154bce-bf2e-4b07-81d9-f2178f71973a" (UID: "ee154bce-bf2e-4b07-81d9-f2178f71973a"). InnerVolumeSpecName "kube-api-access-47lhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.168588 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ee154bce-bf2e-4b07-81d9-f2178f71973a" (UID: "ee154bce-bf2e-4b07-81d9-f2178f71973a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.207963 4696 generic.go:334] "Generic (PLEG): container finished" podID="ee154bce-bf2e-4b07-81d9-f2178f71973a" containerID="56edcf19c7b5e1e663a3e748a2ce752c2f1fcccf530bf4e57a55cdc8db2ffe27" exitCode=0 Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.208029 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee154bce-bf2e-4b07-81d9-f2178f71973a","Type":"ContainerDied","Data":"56edcf19c7b5e1e663a3e748a2ce752c2f1fcccf530bf4e57a55cdc8db2ffe27"} Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.208055 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee154bce-bf2e-4b07-81d9-f2178f71973a","Type":"ContainerDied","Data":"f88c15c41b0306cb7a58f39f784e32361c971beb5b5af33efce45ed4f1a3a785"} Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.208072 4696 scope.go:117] "RemoveContainer" containerID="11b947598c7c8c0c8df9873f444b0ab7bbaa43d2d60fc6a236874da075946586" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.208177 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.209810 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-j298g" event={"ID":"5e448486-5596-4442-a11a-804a7b7b94d8","Type":"ContainerDied","Data":"371f9e37d3154054819d68cf79e88d5ee346e2c14faf64f6d34a041e087dc079"} Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.209893 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="371f9e37d3154054819d68cf79e88d5ee346e2c14faf64f6d34a041e087dc079" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.210304 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-j298g" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.239960 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee154bce-bf2e-4b07-81d9-f2178f71973a" (UID: "ee154bce-bf2e-4b07-81d9-f2178f71973a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.243678 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.243707 4696 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.243716 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47lhx\" (UniqueName: \"kubernetes.io/projected/ee154bce-bf2e-4b07-81d9-f2178f71973a-kube-api-access-47lhx\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.243725 4696 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee154bce-bf2e-4b07-81d9-f2178f71973a-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.243733 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.255476 4696 scope.go:117] "RemoveContainer" containerID="09dc39ceb111d4ebcced1a1a7c1b527465533c6bc4ad8f2d2086bdc85a5b3bf4" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.260758 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-config-data" (OuterVolumeSpecName: "config-data") pod "ee154bce-bf2e-4b07-81d9-f2178f71973a" (UID: "ee154bce-bf2e-4b07-81d9-f2178f71973a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.281035 4696 scope.go:117] "RemoveContainer" containerID="193a3d19134bb486e9301c5545a00e7461f8c3beb00ed97011d47dd42540b7d2" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.312110 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 21 08:49:13 crc kubenswrapper[4696]: E0321 08:49:13.312554 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee154bce-bf2e-4b07-81d9-f2178f71973a" containerName="proxy-httpd" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.312570 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee154bce-bf2e-4b07-81d9-f2178f71973a" containerName="proxy-httpd" Mar 21 08:49:13 crc kubenswrapper[4696]: E0321 08:49:13.312585 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee154bce-bf2e-4b07-81d9-f2178f71973a" containerName="sg-core" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.312592 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee154bce-bf2e-4b07-81d9-f2178f71973a" containerName="sg-core" Mar 21 08:49:13 crc kubenswrapper[4696]: E0321 08:49:13.312605 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e448486-5596-4442-a11a-804a7b7b94d8" containerName="nova-cell0-conductor-db-sync" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.312612 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e448486-5596-4442-a11a-804a7b7b94d8" containerName="nova-cell0-conductor-db-sync" Mar 21 08:49:13 crc kubenswrapper[4696]: E0321 08:49:13.312624 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee154bce-bf2e-4b07-81d9-f2178f71973a" containerName="ceilometer-notification-agent" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.312630 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee154bce-bf2e-4b07-81d9-f2178f71973a" containerName="ceilometer-notification-agent" Mar 21 08:49:13 crc kubenswrapper[4696]: E0321 08:49:13.312647 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee154bce-bf2e-4b07-81d9-f2178f71973a" containerName="ceilometer-central-agent" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.312652 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee154bce-bf2e-4b07-81d9-f2178f71973a" containerName="ceilometer-central-agent" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.312847 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee154bce-bf2e-4b07-81d9-f2178f71973a" containerName="proxy-httpd" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.312870 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e448486-5596-4442-a11a-804a7b7b94d8" containerName="nova-cell0-conductor-db-sync" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.312881 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee154bce-bf2e-4b07-81d9-f2178f71973a" containerName="sg-core" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.312888 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee154bce-bf2e-4b07-81d9-f2178f71973a" containerName="ceilometer-notification-agent" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.312907 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee154bce-bf2e-4b07-81d9-f2178f71973a" containerName="ceilometer-central-agent" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.313704 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.316619 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-psnkw" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.316878 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.324595 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.327565 4696 scope.go:117] "RemoveContainer" containerID="56edcf19c7b5e1e663a3e748a2ce752c2f1fcccf530bf4e57a55cdc8db2ffe27" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.349115 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba86395-973a-4d8b-976f-673403ae58e5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7ba86395-973a-4d8b-976f-673403ae58e5\") " pod="openstack/nova-cell0-conductor-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.349192 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pj9jc\" (UniqueName: \"kubernetes.io/projected/7ba86395-973a-4d8b-976f-673403ae58e5-kube-api-access-pj9jc\") pod \"nova-cell0-conductor-0\" (UID: \"7ba86395-973a-4d8b-976f-673403ae58e5\") " pod="openstack/nova-cell0-conductor-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.349286 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ba86395-973a-4d8b-976f-673403ae58e5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7ba86395-973a-4d8b-976f-673403ae58e5\") " pod="openstack/nova-cell0-conductor-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.349413 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee154bce-bf2e-4b07-81d9-f2178f71973a-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.362600 4696 scope.go:117] "RemoveContainer" containerID="11b947598c7c8c0c8df9873f444b0ab7bbaa43d2d60fc6a236874da075946586" Mar 21 08:49:13 crc kubenswrapper[4696]: E0321 08:49:13.363590 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11b947598c7c8c0c8df9873f444b0ab7bbaa43d2d60fc6a236874da075946586\": container with ID starting with 11b947598c7c8c0c8df9873f444b0ab7bbaa43d2d60fc6a236874da075946586 not found: ID does not exist" containerID="11b947598c7c8c0c8df9873f444b0ab7bbaa43d2d60fc6a236874da075946586" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.363618 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11b947598c7c8c0c8df9873f444b0ab7bbaa43d2d60fc6a236874da075946586"} err="failed to get container status \"11b947598c7c8c0c8df9873f444b0ab7bbaa43d2d60fc6a236874da075946586\": rpc error: code = NotFound desc = could not find container \"11b947598c7c8c0c8df9873f444b0ab7bbaa43d2d60fc6a236874da075946586\": container with ID starting with 11b947598c7c8c0c8df9873f444b0ab7bbaa43d2d60fc6a236874da075946586 not found: ID does not exist" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.363637 4696 scope.go:117] "RemoveContainer" containerID="09dc39ceb111d4ebcced1a1a7c1b527465533c6bc4ad8f2d2086bdc85a5b3bf4" Mar 21 08:49:13 crc kubenswrapper[4696]: E0321 08:49:13.365664 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09dc39ceb111d4ebcced1a1a7c1b527465533c6bc4ad8f2d2086bdc85a5b3bf4\": container with ID starting with 09dc39ceb111d4ebcced1a1a7c1b527465533c6bc4ad8f2d2086bdc85a5b3bf4 not found: ID does not exist" containerID="09dc39ceb111d4ebcced1a1a7c1b527465533c6bc4ad8f2d2086bdc85a5b3bf4" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.365684 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09dc39ceb111d4ebcced1a1a7c1b527465533c6bc4ad8f2d2086bdc85a5b3bf4"} err="failed to get container status \"09dc39ceb111d4ebcced1a1a7c1b527465533c6bc4ad8f2d2086bdc85a5b3bf4\": rpc error: code = NotFound desc = could not find container \"09dc39ceb111d4ebcced1a1a7c1b527465533c6bc4ad8f2d2086bdc85a5b3bf4\": container with ID starting with 09dc39ceb111d4ebcced1a1a7c1b527465533c6bc4ad8f2d2086bdc85a5b3bf4 not found: ID does not exist" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.365700 4696 scope.go:117] "RemoveContainer" containerID="193a3d19134bb486e9301c5545a00e7461f8c3beb00ed97011d47dd42540b7d2" Mar 21 08:49:13 crc kubenswrapper[4696]: E0321 08:49:13.366122 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"193a3d19134bb486e9301c5545a00e7461f8c3beb00ed97011d47dd42540b7d2\": container with ID starting with 193a3d19134bb486e9301c5545a00e7461f8c3beb00ed97011d47dd42540b7d2 not found: ID does not exist" containerID="193a3d19134bb486e9301c5545a00e7461f8c3beb00ed97011d47dd42540b7d2" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.366151 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"193a3d19134bb486e9301c5545a00e7461f8c3beb00ed97011d47dd42540b7d2"} err="failed to get container status \"193a3d19134bb486e9301c5545a00e7461f8c3beb00ed97011d47dd42540b7d2\": rpc error: code = NotFound desc = could not find container \"193a3d19134bb486e9301c5545a00e7461f8c3beb00ed97011d47dd42540b7d2\": container with ID starting with 193a3d19134bb486e9301c5545a00e7461f8c3beb00ed97011d47dd42540b7d2 not found: ID does not exist" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.366167 4696 scope.go:117] "RemoveContainer" containerID="56edcf19c7b5e1e663a3e748a2ce752c2f1fcccf530bf4e57a55cdc8db2ffe27" Mar 21 08:49:13 crc kubenswrapper[4696]: E0321 08:49:13.366473 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56edcf19c7b5e1e663a3e748a2ce752c2f1fcccf530bf4e57a55cdc8db2ffe27\": container with ID starting with 56edcf19c7b5e1e663a3e748a2ce752c2f1fcccf530bf4e57a55cdc8db2ffe27 not found: ID does not exist" containerID="56edcf19c7b5e1e663a3e748a2ce752c2f1fcccf530bf4e57a55cdc8db2ffe27" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.366497 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56edcf19c7b5e1e663a3e748a2ce752c2f1fcccf530bf4e57a55cdc8db2ffe27"} err="failed to get container status \"56edcf19c7b5e1e663a3e748a2ce752c2f1fcccf530bf4e57a55cdc8db2ffe27\": rpc error: code = NotFound desc = could not find container \"56edcf19c7b5e1e663a3e748a2ce752c2f1fcccf530bf4e57a55cdc8db2ffe27\": container with ID starting with 56edcf19c7b5e1e663a3e748a2ce752c2f1fcccf530bf4e57a55cdc8db2ffe27 not found: ID does not exist" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.451106 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba86395-973a-4d8b-976f-673403ae58e5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7ba86395-973a-4d8b-976f-673403ae58e5\") " pod="openstack/nova-cell0-conductor-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.451174 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pj9jc\" (UniqueName: \"kubernetes.io/projected/7ba86395-973a-4d8b-976f-673403ae58e5-kube-api-access-pj9jc\") pod \"nova-cell0-conductor-0\" (UID: \"7ba86395-973a-4d8b-976f-673403ae58e5\") " pod="openstack/nova-cell0-conductor-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.451260 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ba86395-973a-4d8b-976f-673403ae58e5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7ba86395-973a-4d8b-976f-673403ae58e5\") " pod="openstack/nova-cell0-conductor-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.454446 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ba86395-973a-4d8b-976f-673403ae58e5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7ba86395-973a-4d8b-976f-673403ae58e5\") " pod="openstack/nova-cell0-conductor-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.454505 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba86395-973a-4d8b-976f-673403ae58e5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7ba86395-973a-4d8b-976f-673403ae58e5\") " pod="openstack/nova-cell0-conductor-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.469372 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pj9jc\" (UniqueName: \"kubernetes.io/projected/7ba86395-973a-4d8b-976f-673403ae58e5-kube-api-access-pj9jc\") pod \"nova-cell0-conductor-0\" (UID: \"7ba86395-973a-4d8b-976f-673403ae58e5\") " pod="openstack/nova-cell0-conductor-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.545734 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.555014 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.571705 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.574272 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.577076 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.577156 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.586561 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.636571 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.658095 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f9d152a-965e-417f-9d2a-5ee0d411cf12-log-httpd\") pod \"ceilometer-0\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.658158 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f9d152a-965e-417f-9d2a-5ee0d411cf12-run-httpd\") pod \"ceilometer-0\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.658209 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.658262 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-scripts\") pod \"ceilometer-0\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.658310 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-config-data\") pod \"ceilometer-0\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.658350 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.658389 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-949tz\" (UniqueName: \"kubernetes.io/projected/8f9d152a-965e-417f-9d2a-5ee0d411cf12-kube-api-access-949tz\") pod \"ceilometer-0\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.760294 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-config-data\") pod \"ceilometer-0\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.760549 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.760585 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-949tz\" (UniqueName: \"kubernetes.io/projected/8f9d152a-965e-417f-9d2a-5ee0d411cf12-kube-api-access-949tz\") pod \"ceilometer-0\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.760652 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f9d152a-965e-417f-9d2a-5ee0d411cf12-log-httpd\") pod \"ceilometer-0\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.760683 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f9d152a-965e-417f-9d2a-5ee0d411cf12-run-httpd\") pod \"ceilometer-0\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.760720 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.760762 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-scripts\") pod \"ceilometer-0\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.761293 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f9d152a-965e-417f-9d2a-5ee0d411cf12-log-httpd\") pod \"ceilometer-0\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.761772 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f9d152a-965e-417f-9d2a-5ee0d411cf12-run-httpd\") pod \"ceilometer-0\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.768649 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.768690 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-scripts\") pod \"ceilometer-0\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.769353 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-config-data\") pod \"ceilometer-0\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.772420 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.781099 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-949tz\" (UniqueName: \"kubernetes.io/projected/8f9d152a-965e-417f-9d2a-5ee0d411cf12-kube-api-access-949tz\") pod \"ceilometer-0\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " pod="openstack/ceilometer-0" Mar 21 08:49:13 crc kubenswrapper[4696]: I0321 08:49:13.947407 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:49:14 crc kubenswrapper[4696]: I0321 08:49:14.115314 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 21 08:49:14 crc kubenswrapper[4696]: I0321 08:49:14.233093 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7ba86395-973a-4d8b-976f-673403ae58e5","Type":"ContainerStarted","Data":"7c31f3120ff897e765f1495094d105d55b714e85ad963e0bb90c0cac07de599f"} Mar 21 08:49:14 crc kubenswrapper[4696]: W0321 08:49:14.473657 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f9d152a_965e_417f_9d2a_5ee0d411cf12.slice/crio-14f6dbe7980776a050a76fa8aabb1eff2f38e4f58719c2bebccd88b1986047f6 WatchSource:0}: Error finding container 14f6dbe7980776a050a76fa8aabb1eff2f38e4f58719c2bebccd88b1986047f6: Status 404 returned error can't find the container with id 14f6dbe7980776a050a76fa8aabb1eff2f38e4f58719c2bebccd88b1986047f6 Mar 21 08:49:14 crc kubenswrapper[4696]: I0321 08:49:14.477348 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:49:14 crc kubenswrapper[4696]: I0321 08:49:14.521375 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 21 08:49:14 crc kubenswrapper[4696]: I0321 08:49:14.521653 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 21 08:49:14 crc kubenswrapper[4696]: I0321 08:49:14.550328 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee154bce-bf2e-4b07-81d9-f2178f71973a" path="/var/lib/kubelet/pods/ee154bce-bf2e-4b07-81d9-f2178f71973a/volumes" Mar 21 08:49:14 crc kubenswrapper[4696]: I0321 08:49:14.558255 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 21 08:49:14 crc kubenswrapper[4696]: I0321 08:49:14.564349 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 21 08:49:15 crc kubenswrapper[4696]: I0321 08:49:15.244739 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f9d152a-965e-417f-9d2a-5ee0d411cf12","Type":"ContainerStarted","Data":"38ccce59ac2ffe1ae38f78152225c4bfb67b306e98a929c1fed3c61e3da1fc33"} Mar 21 08:49:15 crc kubenswrapper[4696]: I0321 08:49:15.244783 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f9d152a-965e-417f-9d2a-5ee0d411cf12","Type":"ContainerStarted","Data":"14f6dbe7980776a050a76fa8aabb1eff2f38e4f58719c2bebccd88b1986047f6"} Mar 21 08:49:15 crc kubenswrapper[4696]: I0321 08:49:15.250066 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7ba86395-973a-4d8b-976f-673403ae58e5","Type":"ContainerStarted","Data":"b6c5db51f8ae5d3cf25dd4f4753fe608930b233282b9d30881b0728f907a7f15"} Mar 21 08:49:15 crc kubenswrapper[4696]: I0321 08:49:15.250148 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 21 08:49:15 crc kubenswrapper[4696]: I0321 08:49:15.250170 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 21 08:49:15 crc kubenswrapper[4696]: I0321 08:49:15.250185 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 21 08:49:15 crc kubenswrapper[4696]: I0321 08:49:15.271985 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.271965078 podStartE2EDuration="2.271965078s" podCreationTimestamp="2026-03-21 08:49:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:49:15.264230705 +0000 UTC m=+1289.385111418" watchObservedRunningTime="2026-03-21 08:49:15.271965078 +0000 UTC m=+1289.392845791" Mar 21 08:49:16 crc kubenswrapper[4696]: I0321 08:49:16.264074 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f9d152a-965e-417f-9d2a-5ee0d411cf12","Type":"ContainerStarted","Data":"605519f7f48e5626717d9a929e3e972613875f19f9f9565786164fb969ea5533"} Mar 21 08:49:16 crc kubenswrapper[4696]: I0321 08:49:16.825972 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 21 08:49:16 crc kubenswrapper[4696]: I0321 08:49:16.826551 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 21 08:49:16 crc kubenswrapper[4696]: I0321 08:49:16.917080 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 21 08:49:16 crc kubenswrapper[4696]: I0321 08:49:16.952702 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 21 08:49:17 crc kubenswrapper[4696]: I0321 08:49:17.292909 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f9d152a-965e-417f-9d2a-5ee0d411cf12","Type":"ContainerStarted","Data":"ce642d451ff7b23eac072a6c4b9fe87917a4e4360fb55fc439c5a56c6b83b292"} Mar 21 08:49:17 crc kubenswrapper[4696]: I0321 08:49:17.295094 4696 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 21 08:49:17 crc kubenswrapper[4696]: I0321 08:49:17.295111 4696 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 21 08:49:17 crc kubenswrapper[4696]: I0321 08:49:17.296157 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 21 08:49:17 crc kubenswrapper[4696]: I0321 08:49:17.296179 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 21 08:49:18 crc kubenswrapper[4696]: I0321 08:49:18.545755 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 21 08:49:18 crc kubenswrapper[4696]: I0321 08:49:18.546111 4696 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 21 08:49:18 crc kubenswrapper[4696]: I0321 08:49:18.632547 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 21 08:49:19 crc kubenswrapper[4696]: I0321 08:49:19.316680 4696 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 21 08:49:19 crc kubenswrapper[4696]: I0321 08:49:19.316702 4696 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 21 08:49:19 crc kubenswrapper[4696]: I0321 08:49:19.316961 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f9d152a-965e-417f-9d2a-5ee0d411cf12","Type":"ContainerStarted","Data":"626adfe6cca1656f5fa21455eb756144d6df2a606159ca610ad2796c6c8b6054"} Mar 21 08:49:19 crc kubenswrapper[4696]: I0321 08:49:19.317102 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 21 08:49:19 crc kubenswrapper[4696]: I0321 08:49:19.962549 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 21 08:49:19 crc kubenswrapper[4696]: I0321 08:49:19.983135 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.14554436 podStartE2EDuration="6.983118109s" podCreationTimestamp="2026-03-21 08:49:13 +0000 UTC" firstStartedPulling="2026-03-21 08:49:14.475613136 +0000 UTC m=+1288.596493849" lastFinishedPulling="2026-03-21 08:49:18.313186885 +0000 UTC m=+1292.434067598" observedRunningTime="2026-03-21 08:49:19.349562014 +0000 UTC m=+1293.470442727" watchObservedRunningTime="2026-03-21 08:49:19.983118109 +0000 UTC m=+1294.103998822" Mar 21 08:49:20 crc kubenswrapper[4696]: I0321 08:49:20.080553 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 21 08:49:23 crc kubenswrapper[4696]: I0321 08:49:23.457808 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:49:23 crc kubenswrapper[4696]: I0321 08:49:23.458599 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8f9d152a-965e-417f-9d2a-5ee0d411cf12" containerName="ceilometer-central-agent" containerID="cri-o://38ccce59ac2ffe1ae38f78152225c4bfb67b306e98a929c1fed3c61e3da1fc33" gracePeriod=30 Mar 21 08:49:23 crc kubenswrapper[4696]: I0321 08:49:23.459106 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8f9d152a-965e-417f-9d2a-5ee0d411cf12" containerName="sg-core" containerID="cri-o://ce642d451ff7b23eac072a6c4b9fe87917a4e4360fb55fc439c5a56c6b83b292" gracePeriod=30 Mar 21 08:49:23 crc kubenswrapper[4696]: I0321 08:49:23.459163 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8f9d152a-965e-417f-9d2a-5ee0d411cf12" containerName="ceilometer-notification-agent" containerID="cri-o://605519f7f48e5626717d9a929e3e972613875f19f9f9565786164fb969ea5533" gracePeriod=30 Mar 21 08:49:23 crc kubenswrapper[4696]: I0321 08:49:23.459230 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8f9d152a-965e-417f-9d2a-5ee0d411cf12" containerName="proxy-httpd" containerID="cri-o://626adfe6cca1656f5fa21455eb756144d6df2a606159ca610ad2796c6c8b6054" gracePeriod=30 Mar 21 08:49:23 crc kubenswrapper[4696]: I0321 08:49:23.675286 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.143456 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-d4v88"] Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.144727 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-d4v88" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.146695 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.153931 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.160028 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-d4v88"] Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.284593 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-scripts\") pod \"nova-cell0-cell-mapping-d4v88\" (UID: \"29e6fe18-05b8-430f-b3fe-96ca1cf04a94\") " pod="openstack/nova-cell0-cell-mapping-d4v88" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.284660 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-d4v88\" (UID: \"29e6fe18-05b8-430f-b3fe-96ca1cf04a94\") " pod="openstack/nova-cell0-cell-mapping-d4v88" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.284758 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-config-data\") pod \"nova-cell0-cell-mapping-d4v88\" (UID: \"29e6fe18-05b8-430f-b3fe-96ca1cf04a94\") " pod="openstack/nova-cell0-cell-mapping-d4v88" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.285182 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zncj4\" (UniqueName: \"kubernetes.io/projected/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-kube-api-access-zncj4\") pod \"nova-cell0-cell-mapping-d4v88\" (UID: \"29e6fe18-05b8-430f-b3fe-96ca1cf04a94\") " pod="openstack/nova-cell0-cell-mapping-d4v88" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.341364 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.359174 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.359299 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.362115 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.378453 4696 generic.go:334] "Generic (PLEG): container finished" podID="8f9d152a-965e-417f-9d2a-5ee0d411cf12" containerID="626adfe6cca1656f5fa21455eb756144d6df2a606159ca610ad2796c6c8b6054" exitCode=0 Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.378497 4696 generic.go:334] "Generic (PLEG): container finished" podID="8f9d152a-965e-417f-9d2a-5ee0d411cf12" containerID="ce642d451ff7b23eac072a6c4b9fe87917a4e4360fb55fc439c5a56c6b83b292" exitCode=2 Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.378509 4696 generic.go:334] "Generic (PLEG): container finished" podID="8f9d152a-965e-417f-9d2a-5ee0d411cf12" containerID="605519f7f48e5626717d9a929e3e972613875f19f9f9565786164fb969ea5533" exitCode=0 Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.378531 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f9d152a-965e-417f-9d2a-5ee0d411cf12","Type":"ContainerDied","Data":"626adfe6cca1656f5fa21455eb756144d6df2a606159ca610ad2796c6c8b6054"} Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.378560 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f9d152a-965e-417f-9d2a-5ee0d411cf12","Type":"ContainerDied","Data":"ce642d451ff7b23eac072a6c4b9fe87917a4e4360fb55fc439c5a56c6b83b292"} Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.378589 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f9d152a-965e-417f-9d2a-5ee0d411cf12","Type":"ContainerDied","Data":"605519f7f48e5626717d9a929e3e972613875f19f9f9565786164fb969ea5533"} Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.407431 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9lpt\" (UniqueName: \"kubernetes.io/projected/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-kube-api-access-s9lpt\") pod \"nova-api-0\" (UID: \"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1\") " pod="openstack/nova-api-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.407526 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-scripts\") pod \"nova-cell0-cell-mapping-d4v88\" (UID: \"29e6fe18-05b8-430f-b3fe-96ca1cf04a94\") " pod="openstack/nova-cell0-cell-mapping-d4v88" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.407619 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-d4v88\" (UID: \"29e6fe18-05b8-430f-b3fe-96ca1cf04a94\") " pod="openstack/nova-cell0-cell-mapping-d4v88" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.407754 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-logs\") pod \"nova-api-0\" (UID: \"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1\") " pod="openstack/nova-api-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.407929 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-config-data\") pod \"nova-api-0\" (UID: \"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1\") " pod="openstack/nova-api-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.408024 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1\") " pod="openstack/nova-api-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.408100 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-config-data\") pod \"nova-cell0-cell-mapping-d4v88\" (UID: \"29e6fe18-05b8-430f-b3fe-96ca1cf04a94\") " pod="openstack/nova-cell0-cell-mapping-d4v88" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.408503 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zncj4\" (UniqueName: \"kubernetes.io/projected/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-kube-api-access-zncj4\") pod \"nova-cell0-cell-mapping-d4v88\" (UID: \"29e6fe18-05b8-430f-b3fe-96ca1cf04a94\") " pod="openstack/nova-cell0-cell-mapping-d4v88" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.418044 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.444675 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zncj4\" (UniqueName: \"kubernetes.io/projected/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-kube-api-access-zncj4\") pod \"nova-cell0-cell-mapping-d4v88\" (UID: \"29e6fe18-05b8-430f-b3fe-96ca1cf04a94\") " pod="openstack/nova-cell0-cell-mapping-d4v88" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.449173 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.449942 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-config-data\") pod \"nova-cell0-cell-mapping-d4v88\" (UID: \"29e6fe18-05b8-430f-b3fe-96ca1cf04a94\") " pod="openstack/nova-cell0-cell-mapping-d4v88" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.457778 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-d4v88\" (UID: \"29e6fe18-05b8-430f-b3fe-96ca1cf04a94\") " pod="openstack/nova-cell0-cell-mapping-d4v88" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.460199 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-scripts\") pod \"nova-cell0-cell-mapping-d4v88\" (UID: \"29e6fe18-05b8-430f-b3fe-96ca1cf04a94\") " pod="openstack/nova-cell0-cell-mapping-d4v88" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.460976 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-d4v88" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.466077 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.512409 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.513935 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9lpt\" (UniqueName: \"kubernetes.io/projected/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-kube-api-access-s9lpt\") pod \"nova-api-0\" (UID: \"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1\") " pod="openstack/nova-api-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.514003 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-logs\") pod \"nova-api-0\" (UID: \"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1\") " pod="openstack/nova-api-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.514033 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-config-data\") pod \"nova-api-0\" (UID: \"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1\") " pod="openstack/nova-api-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.514059 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1\") " pod="openstack/nova-api-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.516679 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-logs\") pod \"nova-api-0\" (UID: \"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1\") " pod="openstack/nova-api-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.533261 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1\") " pod="openstack/nova-api-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.554658 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-config-data\") pod \"nova-api-0\" (UID: \"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1\") " pod="openstack/nova-api-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.569419 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9lpt\" (UniqueName: \"kubernetes.io/projected/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-kube-api-access-s9lpt\") pod \"nova-api-0\" (UID: \"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1\") " pod="openstack/nova-api-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.593731 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.594999 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.597145 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.600529 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.616238 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b436237-c742-46e2-a4b0-f0eec79d9fd2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0b436237-c742-46e2-a4b0-f0eec79d9fd2\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.616395 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b436237-c742-46e2-a4b0-f0eec79d9fd2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0b436237-c742-46e2-a4b0-f0eec79d9fd2\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.616417 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwwh9\" (UniqueName: \"kubernetes.io/projected/0b436237-c742-46e2-a4b0-f0eec79d9fd2-kube-api-access-bwwh9\") pod \"nova-cell1-novncproxy-0\" (UID: \"0b436237-c742-46e2-a4b0-f0eec79d9fd2\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.616516 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.630570 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.637416 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.637715 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.652074 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78cd565959-cg8dx"] Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.654005 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.677027 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78cd565959-cg8dx"] Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.688532 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.727788 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae265c6e-25bb-421d-a717-78e42a3e38ca-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ae265c6e-25bb-421d-a717-78e42a3e38ca\") " pod="openstack/nova-metadata-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.728135 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b436237-c742-46e2-a4b0-f0eec79d9fd2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0b436237-c742-46e2-a4b0-f0eec79d9fd2\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.728160 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae265c6e-25bb-421d-a717-78e42a3e38ca-config-data\") pod \"nova-metadata-0\" (UID: \"ae265c6e-25bb-421d-a717-78e42a3e38ca\") " pod="openstack/nova-metadata-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.728196 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9dxq\" (UniqueName: \"kubernetes.io/projected/3f429da6-5739-4444-8b85-347dc971df0f-kube-api-access-l9dxq\") pod \"dnsmasq-dns-78cd565959-cg8dx\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.728301 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a1d264d-7e6b-4910-9656-61ffefe53f8d-config-data\") pod \"nova-scheduler-0\" (UID: \"7a1d264d-7e6b-4910-9656-61ffefe53f8d\") " pod="openstack/nova-scheduler-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.728379 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-config\") pod \"dnsmasq-dns-78cd565959-cg8dx\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.728479 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae265c6e-25bb-421d-a717-78e42a3e38ca-logs\") pod \"nova-metadata-0\" (UID: \"ae265c6e-25bb-421d-a717-78e42a3e38ca\") " pod="openstack/nova-metadata-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.728543 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-dns-svc\") pod \"dnsmasq-dns-78cd565959-cg8dx\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.728602 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4s5z\" (UniqueName: \"kubernetes.io/projected/7a1d264d-7e6b-4910-9656-61ffefe53f8d-kube-api-access-r4s5z\") pod \"nova-scheduler-0\" (UID: \"7a1d264d-7e6b-4910-9656-61ffefe53f8d\") " pod="openstack/nova-scheduler-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.728627 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-dns-swift-storage-0\") pod \"dnsmasq-dns-78cd565959-cg8dx\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.728721 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a1d264d-7e6b-4910-9656-61ffefe53f8d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7a1d264d-7e6b-4910-9656-61ffefe53f8d\") " pod="openstack/nova-scheduler-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.728774 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk4wl\" (UniqueName: \"kubernetes.io/projected/ae265c6e-25bb-421d-a717-78e42a3e38ca-kube-api-access-jk4wl\") pod \"nova-metadata-0\" (UID: \"ae265c6e-25bb-421d-a717-78e42a3e38ca\") " pod="openstack/nova-metadata-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.728914 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-ovsdbserver-sb\") pod \"dnsmasq-dns-78cd565959-cg8dx\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.728949 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b436237-c742-46e2-a4b0-f0eec79d9fd2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0b436237-c742-46e2-a4b0-f0eec79d9fd2\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.728969 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwwh9\" (UniqueName: \"kubernetes.io/projected/0b436237-c742-46e2-a4b0-f0eec79d9fd2-kube-api-access-bwwh9\") pod \"nova-cell1-novncproxy-0\" (UID: \"0b436237-c742-46e2-a4b0-f0eec79d9fd2\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.729533 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-ovsdbserver-nb\") pod \"dnsmasq-dns-78cd565959-cg8dx\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.747641 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b436237-c742-46e2-a4b0-f0eec79d9fd2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0b436237-c742-46e2-a4b0-f0eec79d9fd2\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.751523 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b436237-c742-46e2-a4b0-f0eec79d9fd2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0b436237-c742-46e2-a4b0-f0eec79d9fd2\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.770842 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwwh9\" (UniqueName: \"kubernetes.io/projected/0b436237-c742-46e2-a4b0-f0eec79d9fd2-kube-api-access-bwwh9\") pod \"nova-cell1-novncproxy-0\" (UID: \"0b436237-c742-46e2-a4b0-f0eec79d9fd2\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.834265 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-dns-svc\") pod \"dnsmasq-dns-78cd565959-cg8dx\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.834324 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4s5z\" (UniqueName: \"kubernetes.io/projected/7a1d264d-7e6b-4910-9656-61ffefe53f8d-kube-api-access-r4s5z\") pod \"nova-scheduler-0\" (UID: \"7a1d264d-7e6b-4910-9656-61ffefe53f8d\") " pod="openstack/nova-scheduler-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.834344 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-dns-swift-storage-0\") pod \"dnsmasq-dns-78cd565959-cg8dx\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.834376 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a1d264d-7e6b-4910-9656-61ffefe53f8d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7a1d264d-7e6b-4910-9656-61ffefe53f8d\") " pod="openstack/nova-scheduler-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.834396 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk4wl\" (UniqueName: \"kubernetes.io/projected/ae265c6e-25bb-421d-a717-78e42a3e38ca-kube-api-access-jk4wl\") pod \"nova-metadata-0\" (UID: \"ae265c6e-25bb-421d-a717-78e42a3e38ca\") " pod="openstack/nova-metadata-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.834426 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-ovsdbserver-sb\") pod \"dnsmasq-dns-78cd565959-cg8dx\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.834458 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-ovsdbserver-nb\") pod \"dnsmasq-dns-78cd565959-cg8dx\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.834546 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae265c6e-25bb-421d-a717-78e42a3e38ca-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ae265c6e-25bb-421d-a717-78e42a3e38ca\") " pod="openstack/nova-metadata-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.834592 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae265c6e-25bb-421d-a717-78e42a3e38ca-config-data\") pod \"nova-metadata-0\" (UID: \"ae265c6e-25bb-421d-a717-78e42a3e38ca\") " pod="openstack/nova-metadata-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.834609 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9dxq\" (UniqueName: \"kubernetes.io/projected/3f429da6-5739-4444-8b85-347dc971df0f-kube-api-access-l9dxq\") pod \"dnsmasq-dns-78cd565959-cg8dx\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.834628 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a1d264d-7e6b-4910-9656-61ffefe53f8d-config-data\") pod \"nova-scheduler-0\" (UID: \"7a1d264d-7e6b-4910-9656-61ffefe53f8d\") " pod="openstack/nova-scheduler-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.834685 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-config\") pod \"dnsmasq-dns-78cd565959-cg8dx\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.834739 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae265c6e-25bb-421d-a717-78e42a3e38ca-logs\") pod \"nova-metadata-0\" (UID: \"ae265c6e-25bb-421d-a717-78e42a3e38ca\") " pod="openstack/nova-metadata-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.835329 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae265c6e-25bb-421d-a717-78e42a3e38ca-logs\") pod \"nova-metadata-0\" (UID: \"ae265c6e-25bb-421d-a717-78e42a3e38ca\") " pod="openstack/nova-metadata-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.835798 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-dns-svc\") pod \"dnsmasq-dns-78cd565959-cg8dx\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.839774 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.841142 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-ovsdbserver-sb\") pod \"dnsmasq-dns-78cd565959-cg8dx\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.841727 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-config\") pod \"dnsmasq-dns-78cd565959-cg8dx\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.842674 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-dns-swift-storage-0\") pod \"dnsmasq-dns-78cd565959-cg8dx\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.842923 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-ovsdbserver-nb\") pod \"dnsmasq-dns-78cd565959-cg8dx\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.851793 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae265c6e-25bb-421d-a717-78e42a3e38ca-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ae265c6e-25bb-421d-a717-78e42a3e38ca\") " pod="openstack/nova-metadata-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.852015 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae265c6e-25bb-421d-a717-78e42a3e38ca-config-data\") pod \"nova-metadata-0\" (UID: \"ae265c6e-25bb-421d-a717-78e42a3e38ca\") " pod="openstack/nova-metadata-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.854467 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a1d264d-7e6b-4910-9656-61ffefe53f8d-config-data\") pod \"nova-scheduler-0\" (UID: \"7a1d264d-7e6b-4910-9656-61ffefe53f8d\") " pod="openstack/nova-scheduler-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.857608 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a1d264d-7e6b-4910-9656-61ffefe53f8d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7a1d264d-7e6b-4910-9656-61ffefe53f8d\") " pod="openstack/nova-scheduler-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.862632 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9dxq\" (UniqueName: \"kubernetes.io/projected/3f429da6-5739-4444-8b85-347dc971df0f-kube-api-access-l9dxq\") pod \"dnsmasq-dns-78cd565959-cg8dx\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.875048 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk4wl\" (UniqueName: \"kubernetes.io/projected/ae265c6e-25bb-421d-a717-78e42a3e38ca-kube-api-access-jk4wl\") pod \"nova-metadata-0\" (UID: \"ae265c6e-25bb-421d-a717-78e42a3e38ca\") " pod="openstack/nova-metadata-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.875496 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4s5z\" (UniqueName: \"kubernetes.io/projected/7a1d264d-7e6b-4910-9656-61ffefe53f8d-kube-api-access-r4s5z\") pod \"nova-scheduler-0\" (UID: \"7a1d264d-7e6b-4910-9656-61ffefe53f8d\") " pod="openstack/nova-scheduler-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.970981 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 21 08:49:24 crc kubenswrapper[4696]: I0321 08:49:24.982209 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 08:49:25 crc kubenswrapper[4696]: I0321 08:49:25.005986 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:25 crc kubenswrapper[4696]: I0321 08:49:25.196287 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-d4v88"] Mar 21 08:49:25 crc kubenswrapper[4696]: I0321 08:49:25.430993 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-d4v88" event={"ID":"29e6fe18-05b8-430f-b3fe-96ca1cf04a94","Type":"ContainerStarted","Data":"27f55643d95b74efe15532bc302311a0a1e896bee3df443a4421d813c5214d51"} Mar 21 08:49:25 crc kubenswrapper[4696]: W0321 08:49:25.812542 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b436237_c742_46e2_a4b0_f0eec79d9fd2.slice/crio-4aa2117782eac7109f7fb6352e659bd7322965a6a6b3f0e92fb4a5df746bccf9 WatchSource:0}: Error finding container 4aa2117782eac7109f7fb6352e659bd7322965a6a6b3f0e92fb4a5df746bccf9: Status 404 returned error can't find the container with id 4aa2117782eac7109f7fb6352e659bd7322965a6a6b3f0e92fb4a5df746bccf9 Mar 21 08:49:25 crc kubenswrapper[4696]: I0321 08:49:25.817949 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 21 08:49:25 crc kubenswrapper[4696]: I0321 08:49:25.829697 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.049024 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fjcrs"] Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.050453 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fjcrs" Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.052645 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.053023 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.064887 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fjcrs"] Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.168032 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-scripts\") pod \"nova-cell1-conductor-db-sync-fjcrs\" (UID: \"4856e60e-c40e-4b95-bc64-2f33b8b79e4b\") " pod="openstack/nova-cell1-conductor-db-sync-fjcrs" Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.168080 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-fjcrs\" (UID: \"4856e60e-c40e-4b95-bc64-2f33b8b79e4b\") " pod="openstack/nova-cell1-conductor-db-sync-fjcrs" Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.168119 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-config-data\") pod \"nova-cell1-conductor-db-sync-fjcrs\" (UID: \"4856e60e-c40e-4b95-bc64-2f33b8b79e4b\") " pod="openstack/nova-cell1-conductor-db-sync-fjcrs" Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.168162 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmvwv\" (UniqueName: \"kubernetes.io/projected/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-kube-api-access-xmvwv\") pod \"nova-cell1-conductor-db-sync-fjcrs\" (UID: \"4856e60e-c40e-4b95-bc64-2f33b8b79e4b\") " pod="openstack/nova-cell1-conductor-db-sync-fjcrs" Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.234262 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.275946 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-scripts\") pod \"nova-cell1-conductor-db-sync-fjcrs\" (UID: \"4856e60e-c40e-4b95-bc64-2f33b8b79e4b\") " pod="openstack/nova-cell1-conductor-db-sync-fjcrs" Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.276313 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-fjcrs\" (UID: \"4856e60e-c40e-4b95-bc64-2f33b8b79e4b\") " pod="openstack/nova-cell1-conductor-db-sync-fjcrs" Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.276363 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-config-data\") pod \"nova-cell1-conductor-db-sync-fjcrs\" (UID: \"4856e60e-c40e-4b95-bc64-2f33b8b79e4b\") " pod="openstack/nova-cell1-conductor-db-sync-fjcrs" Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.276412 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmvwv\" (UniqueName: \"kubernetes.io/projected/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-kube-api-access-xmvwv\") pod \"nova-cell1-conductor-db-sync-fjcrs\" (UID: \"4856e60e-c40e-4b95-bc64-2f33b8b79e4b\") " pod="openstack/nova-cell1-conductor-db-sync-fjcrs" Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.286409 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-config-data\") pod \"nova-cell1-conductor-db-sync-fjcrs\" (UID: \"4856e60e-c40e-4b95-bc64-2f33b8b79e4b\") " pod="openstack/nova-cell1-conductor-db-sync-fjcrs" Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.289292 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-scripts\") pod \"nova-cell1-conductor-db-sync-fjcrs\" (UID: \"4856e60e-c40e-4b95-bc64-2f33b8b79e4b\") " pod="openstack/nova-cell1-conductor-db-sync-fjcrs" Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.293444 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-fjcrs\" (UID: \"4856e60e-c40e-4b95-bc64-2f33b8b79e4b\") " pod="openstack/nova-cell1-conductor-db-sync-fjcrs" Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.298385 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmvwv\" (UniqueName: \"kubernetes.io/projected/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-kube-api-access-xmvwv\") pod \"nova-cell1-conductor-db-sync-fjcrs\" (UID: \"4856e60e-c40e-4b95-bc64-2f33b8b79e4b\") " pod="openstack/nova-cell1-conductor-db-sync-fjcrs" Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.409396 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78cd565959-cg8dx"] Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.416971 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fjcrs" Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.423693 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 08:49:26 crc kubenswrapper[4696]: W0321 08:49:26.426282 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f429da6_5739_4444_8b85_347dc971df0f.slice/crio-3adf528c84939fa36d51ea254f38bb9553a43b017ea883c60b489b02f5eb9e96 WatchSource:0}: Error finding container 3adf528c84939fa36d51ea254f38bb9553a43b017ea883c60b489b02f5eb9e96: Status 404 returned error can't find the container with id 3adf528c84939fa36d51ea254f38bb9553a43b017ea883c60b489b02f5eb9e96 Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.465085 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7a1d264d-7e6b-4910-9656-61ffefe53f8d","Type":"ContainerStarted","Data":"6fc482dfefa52fc80cb91ce841922647c1ee5e44e9d07ec623c7fadf7e6e19a4"} Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.471041 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78cd565959-cg8dx" event={"ID":"3f429da6-5739-4444-8b85-347dc971df0f","Type":"ContainerStarted","Data":"3adf528c84939fa36d51ea254f38bb9553a43b017ea883c60b489b02f5eb9e96"} Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.473304 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0b436237-c742-46e2-a4b0-f0eec79d9fd2","Type":"ContainerStarted","Data":"4aa2117782eac7109f7fb6352e659bd7322965a6a6b3f0e92fb4a5df746bccf9"} Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.476503 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-d4v88" event={"ID":"29e6fe18-05b8-430f-b3fe-96ca1cf04a94","Type":"ContainerStarted","Data":"32c9d7bf028548a11c13896a74b2d7e729721772af461048e7575f7de03057ed"} Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.482297 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1","Type":"ContainerStarted","Data":"7cbc7eeb169246f38abd7ccd436eee26e9fd9cf70f9f5628d8f6031403cf0f9c"} Mar 21 08:49:26 crc kubenswrapper[4696]: I0321 08:49:26.530158 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-d4v88" podStartSLOduration=2.530140697 podStartE2EDuration="2.530140697s" podCreationTimestamp="2026-03-21 08:49:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:49:26.500451375 +0000 UTC m=+1300.621332088" watchObservedRunningTime="2026-03-21 08:49:26.530140697 +0000 UTC m=+1300.651021410" Mar 21 08:49:27 crc kubenswrapper[4696]: I0321 08:49:27.031218 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fjcrs"] Mar 21 08:49:27 crc kubenswrapper[4696]: W0321 08:49:27.045970 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4856e60e_c40e_4b95_bc64_2f33b8b79e4b.slice/crio-09b52a120a53a4ed65c3df82da374baa2820634eaaff50e517bf1a3719cd1381 WatchSource:0}: Error finding container 09b52a120a53a4ed65c3df82da374baa2820634eaaff50e517bf1a3719cd1381: Status 404 returned error can't find the container with id 09b52a120a53a4ed65c3df82da374baa2820634eaaff50e517bf1a3719cd1381 Mar 21 08:49:27 crc kubenswrapper[4696]: I0321 08:49:27.503863 4696 generic.go:334] "Generic (PLEG): container finished" podID="3f429da6-5739-4444-8b85-347dc971df0f" containerID="37130251119ee5a8480000c6dc37701adcac361a80e5141bae0b5b2c6e4fcf13" exitCode=0 Mar 21 08:49:27 crc kubenswrapper[4696]: I0321 08:49:27.504489 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78cd565959-cg8dx" event={"ID":"3f429da6-5739-4444-8b85-347dc971df0f","Type":"ContainerDied","Data":"37130251119ee5a8480000c6dc37701adcac361a80e5141bae0b5b2c6e4fcf13"} Mar 21 08:49:27 crc kubenswrapper[4696]: I0321 08:49:27.506502 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ae265c6e-25bb-421d-a717-78e42a3e38ca","Type":"ContainerStarted","Data":"c1e1cf21e70d52111c086ff7b02fedf9b74d167eb95439180d4f7addbe589e22"} Mar 21 08:49:27 crc kubenswrapper[4696]: I0321 08:49:27.520181 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fjcrs" event={"ID":"4856e60e-c40e-4b95-bc64-2f33b8b79e4b","Type":"ContainerStarted","Data":"4be79300257c4c766e3f6f08ee86a69724e964c534ae69af90e17eeac61e8aa6"} Mar 21 08:49:27 crc kubenswrapper[4696]: I0321 08:49:27.520226 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fjcrs" event={"ID":"4856e60e-c40e-4b95-bc64-2f33b8b79e4b","Type":"ContainerStarted","Data":"09b52a120a53a4ed65c3df82da374baa2820634eaaff50e517bf1a3719cd1381"} Mar 21 08:49:27 crc kubenswrapper[4696]: I0321 08:49:27.557288 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-fjcrs" podStartSLOduration=1.557269459 podStartE2EDuration="1.557269459s" podCreationTimestamp="2026-03-21 08:49:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:49:27.542286045 +0000 UTC m=+1301.663166758" watchObservedRunningTime="2026-03-21 08:49:27.557269459 +0000 UTC m=+1301.678150162" Mar 21 08:49:27 crc kubenswrapper[4696]: I0321 08:49:27.967584 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 08:49:27 crc kubenswrapper[4696]: I0321 08:49:27.992091 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 21 08:49:28 crc kubenswrapper[4696]: I0321 08:49:28.611094 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78cd565959-cg8dx" event={"ID":"3f429da6-5739-4444-8b85-347dc971df0f","Type":"ContainerStarted","Data":"7cb8c4ea50917367d7fc916a934a75bbe68e00c5926ac97cdd8289edf3d76842"} Mar 21 08:49:28 crc kubenswrapper[4696]: I0321 08:49:28.612910 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:28 crc kubenswrapper[4696]: I0321 08:49:28.646825 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78cd565959-cg8dx" podStartSLOduration=4.646794401 podStartE2EDuration="4.646794401s" podCreationTimestamp="2026-03-21 08:49:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:49:28.633169364 +0000 UTC m=+1302.754050077" watchObservedRunningTime="2026-03-21 08:49:28.646794401 +0000 UTC m=+1302.767675114" Mar 21 08:49:31 crc kubenswrapper[4696]: I0321 08:49:31.658126 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ae265c6e-25bb-421d-a717-78e42a3e38ca","Type":"ContainerStarted","Data":"345e7ef002f347d2d9f10204eacfad5d403c1e573b451455e2599cb2a7224dd8"} Mar 21 08:49:31 crc kubenswrapper[4696]: I0321 08:49:31.661367 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0b436237-c742-46e2-a4b0-f0eec79d9fd2","Type":"ContainerStarted","Data":"68a26d8e09f41c2c1a6c42fa9af0f87c665effab807593d28764e1852ec9d521"} Mar 21 08:49:31 crc kubenswrapper[4696]: I0321 08:49:31.661489 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="0b436237-c742-46e2-a4b0-f0eec79d9fd2" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://68a26d8e09f41c2c1a6c42fa9af0f87c665effab807593d28764e1852ec9d521" gracePeriod=30 Mar 21 08:49:31 crc kubenswrapper[4696]: I0321 08:49:31.670140 4696 generic.go:334] "Generic (PLEG): container finished" podID="8f9d152a-965e-417f-9d2a-5ee0d411cf12" containerID="38ccce59ac2ffe1ae38f78152225c4bfb67b306e98a929c1fed3c61e3da1fc33" exitCode=0 Mar 21 08:49:31 crc kubenswrapper[4696]: I0321 08:49:31.670203 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f9d152a-965e-417f-9d2a-5ee0d411cf12","Type":"ContainerDied","Data":"38ccce59ac2ffe1ae38f78152225c4bfb67b306e98a929c1fed3c61e3da1fc33"} Mar 21 08:49:31 crc kubenswrapper[4696]: I0321 08:49:31.682828 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.450364787 podStartE2EDuration="7.682802523s" podCreationTimestamp="2026-03-21 08:49:24 +0000 UTC" firstStartedPulling="2026-03-21 08:49:25.82101586 +0000 UTC m=+1299.941896573" lastFinishedPulling="2026-03-21 08:49:31.053453596 +0000 UTC m=+1305.174334309" observedRunningTime="2026-03-21 08:49:31.681566629 +0000 UTC m=+1305.802447352" watchObservedRunningTime="2026-03-21 08:49:31.682802523 +0000 UTC m=+1305.803683236" Mar 21 08:49:31 crc kubenswrapper[4696]: I0321 08:49:31.689116 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1","Type":"ContainerStarted","Data":"a5078c0880aedda106502b88c3ed962c04f889b9117d6af7d22c03cccf1fc203"} Mar 21 08:49:31 crc kubenswrapper[4696]: I0321 08:49:31.699876 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7a1d264d-7e6b-4910-9656-61ffefe53f8d","Type":"ContainerStarted","Data":"92ad6b8cfc9519d7f9031eec9efa261d2fe191665ac15c9fca809d3ecc7e6279"} Mar 21 08:49:31 crc kubenswrapper[4696]: I0321 08:49:31.729107 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.93072378 podStartE2EDuration="7.729081485s" podCreationTimestamp="2026-03-21 08:49:24 +0000 UTC" firstStartedPulling="2026-03-21 08:49:26.255576184 +0000 UTC m=+1300.376456897" lastFinishedPulling="2026-03-21 08:49:31.053933889 +0000 UTC m=+1305.174814602" observedRunningTime="2026-03-21 08:49:31.717079122 +0000 UTC m=+1305.837959835" watchObservedRunningTime="2026-03-21 08:49:31.729081485 +0000 UTC m=+1305.849962198" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.173266 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.341715 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-949tz\" (UniqueName: \"kubernetes.io/projected/8f9d152a-965e-417f-9d2a-5ee0d411cf12-kube-api-access-949tz\") pod \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.342106 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-combined-ca-bundle\") pod \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.342141 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f9d152a-965e-417f-9d2a-5ee0d411cf12-log-httpd\") pod \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.342625 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-sg-core-conf-yaml\") pod \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.342487 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f9d152a-965e-417f-9d2a-5ee0d411cf12-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8f9d152a-965e-417f-9d2a-5ee0d411cf12" (UID: "8f9d152a-965e-417f-9d2a-5ee0d411cf12"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.342837 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-config-data\") pod \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.342859 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-scripts\") pod \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.343113 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f9d152a-965e-417f-9d2a-5ee0d411cf12-run-httpd\") pod \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\" (UID: \"8f9d152a-965e-417f-9d2a-5ee0d411cf12\") " Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.343600 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f9d152a-965e-417f-9d2a-5ee0d411cf12-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8f9d152a-965e-417f-9d2a-5ee0d411cf12" (UID: "8f9d152a-965e-417f-9d2a-5ee0d411cf12"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.343735 4696 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f9d152a-965e-417f-9d2a-5ee0d411cf12-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.343746 4696 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f9d152a-965e-417f-9d2a-5ee0d411cf12-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.347926 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-scripts" (OuterVolumeSpecName: "scripts") pod "8f9d152a-965e-417f-9d2a-5ee0d411cf12" (UID: "8f9d152a-965e-417f-9d2a-5ee0d411cf12"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.370799 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f9d152a-965e-417f-9d2a-5ee0d411cf12-kube-api-access-949tz" (OuterVolumeSpecName: "kube-api-access-949tz") pod "8f9d152a-965e-417f-9d2a-5ee0d411cf12" (UID: "8f9d152a-965e-417f-9d2a-5ee0d411cf12"). InnerVolumeSpecName "kube-api-access-949tz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.395331 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8f9d152a-965e-417f-9d2a-5ee0d411cf12" (UID: "8f9d152a-965e-417f-9d2a-5ee0d411cf12"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.445793 4696 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.445881 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.445890 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-949tz\" (UniqueName: \"kubernetes.io/projected/8f9d152a-965e-417f-9d2a-5ee0d411cf12-kube-api-access-949tz\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.466901 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f9d152a-965e-417f-9d2a-5ee0d411cf12" (UID: "8f9d152a-965e-417f-9d2a-5ee0d411cf12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.513184 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-config-data" (OuterVolumeSpecName: "config-data") pod "8f9d152a-965e-417f-9d2a-5ee0d411cf12" (UID: "8f9d152a-965e-417f-9d2a-5ee0d411cf12"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.566655 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.566689 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f9d152a-965e-417f-9d2a-5ee0d411cf12-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.713005 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f9d152a-965e-417f-9d2a-5ee0d411cf12","Type":"ContainerDied","Data":"14f6dbe7980776a050a76fa8aabb1eff2f38e4f58719c2bebccd88b1986047f6"} Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.713059 4696 scope.go:117] "RemoveContainer" containerID="626adfe6cca1656f5fa21455eb756144d6df2a606159ca610ad2796c6c8b6054" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.713199 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.716966 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1","Type":"ContainerStarted","Data":"73da593316c9714471379b39652eebf6cf072a108dd7bbee42f1ef069967be5f"} Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.720092 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ae265c6e-25bb-421d-a717-78e42a3e38ca","Type":"ContainerStarted","Data":"c60e3de586d0628a13c96ac8a17d5a9162547031d23b7cd57a36307afbc59d61"} Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.721144 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ae265c6e-25bb-421d-a717-78e42a3e38ca" containerName="nova-metadata-log" containerID="cri-o://345e7ef002f347d2d9f10204eacfad5d403c1e573b451455e2599cb2a7224dd8" gracePeriod=30 Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.721297 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ae265c6e-25bb-421d-a717-78e42a3e38ca" containerName="nova-metadata-metadata" containerID="cri-o://c60e3de586d0628a13c96ac8a17d5a9162547031d23b7cd57a36307afbc59d61" gracePeriod=30 Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.740117 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.155331022 podStartE2EDuration="8.740098301s" podCreationTimestamp="2026-03-21 08:49:24 +0000 UTC" firstStartedPulling="2026-03-21 08:49:26.476596075 +0000 UTC m=+1300.597476788" lastFinishedPulling="2026-03-21 08:49:31.061363354 +0000 UTC m=+1305.182244067" observedRunningTime="2026-03-21 08:49:32.739154375 +0000 UTC m=+1306.860035088" watchObservedRunningTime="2026-03-21 08:49:32.740098301 +0000 UTC m=+1306.860979024" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.743881 4696 scope.go:117] "RemoveContainer" containerID="ce642d451ff7b23eac072a6c4b9fe87917a4e4360fb55fc439c5a56c6b83b292" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.766387 4696 scope.go:117] "RemoveContainer" containerID="605519f7f48e5626717d9a929e3e972613875f19f9f9565786164fb969ea5533" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.772884 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.785997 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.802961 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:49:32 crc kubenswrapper[4696]: E0321 08:49:32.803506 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f9d152a-965e-417f-9d2a-5ee0d411cf12" containerName="ceilometer-notification-agent" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.803527 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f9d152a-965e-417f-9d2a-5ee0d411cf12" containerName="ceilometer-notification-agent" Mar 21 08:49:32 crc kubenswrapper[4696]: E0321 08:49:32.803555 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f9d152a-965e-417f-9d2a-5ee0d411cf12" containerName="ceilometer-central-agent" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.803561 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f9d152a-965e-417f-9d2a-5ee0d411cf12" containerName="ceilometer-central-agent" Mar 21 08:49:32 crc kubenswrapper[4696]: E0321 08:49:32.803578 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f9d152a-965e-417f-9d2a-5ee0d411cf12" containerName="sg-core" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.803584 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f9d152a-965e-417f-9d2a-5ee0d411cf12" containerName="sg-core" Mar 21 08:49:32 crc kubenswrapper[4696]: E0321 08:49:32.803608 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f9d152a-965e-417f-9d2a-5ee0d411cf12" containerName="proxy-httpd" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.803614 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f9d152a-965e-417f-9d2a-5ee0d411cf12" containerName="proxy-httpd" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.803798 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f9d152a-965e-417f-9d2a-5ee0d411cf12" containerName="sg-core" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.803849 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f9d152a-965e-417f-9d2a-5ee0d411cf12" containerName="proxy-httpd" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.803865 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f9d152a-965e-417f-9d2a-5ee0d411cf12" containerName="ceilometer-central-agent" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.803877 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f9d152a-965e-417f-9d2a-5ee0d411cf12" containerName="ceilometer-notification-agent" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.806164 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.808853 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.809085 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.810443 4696 scope.go:117] "RemoveContainer" containerID="38ccce59ac2ffe1ae38f78152225c4bfb67b306e98a929c1fed3c61e3da1fc33" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.819429 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.594499641 podStartE2EDuration="8.819406138s" podCreationTimestamp="2026-03-21 08:49:24 +0000 UTC" firstStartedPulling="2026-03-21 08:49:25.829381161 +0000 UTC m=+1299.950261874" lastFinishedPulling="2026-03-21 08:49:31.054287658 +0000 UTC m=+1305.175168371" observedRunningTime="2026-03-21 08:49:32.768293982 +0000 UTC m=+1306.889174695" watchObservedRunningTime="2026-03-21 08:49:32.819406138 +0000 UTC m=+1306.940286851" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.846871 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.974897 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " pod="openstack/ceilometer-0" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.974940 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-config-data\") pod \"ceilometer-0\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " pod="openstack/ceilometer-0" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.974975 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-scripts\") pod \"ceilometer-0\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " pod="openstack/ceilometer-0" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.975053 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1df5e20-2002-4b15-a6b6-f23874343914-log-httpd\") pod \"ceilometer-0\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " pod="openstack/ceilometer-0" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.975086 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " pod="openstack/ceilometer-0" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.975121 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfzld\" (UniqueName: \"kubernetes.io/projected/e1df5e20-2002-4b15-a6b6-f23874343914-kube-api-access-bfzld\") pod \"ceilometer-0\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " pod="openstack/ceilometer-0" Mar 21 08:49:32 crc kubenswrapper[4696]: I0321 08:49:32.975152 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1df5e20-2002-4b15-a6b6-f23874343914-run-httpd\") pod \"ceilometer-0\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " pod="openstack/ceilometer-0" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.077196 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-scripts\") pod \"ceilometer-0\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " pod="openstack/ceilometer-0" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.077483 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1df5e20-2002-4b15-a6b6-f23874343914-log-httpd\") pod \"ceilometer-0\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " pod="openstack/ceilometer-0" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.077514 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " pod="openstack/ceilometer-0" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.077548 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfzld\" (UniqueName: \"kubernetes.io/projected/e1df5e20-2002-4b15-a6b6-f23874343914-kube-api-access-bfzld\") pod \"ceilometer-0\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " pod="openstack/ceilometer-0" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.077582 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1df5e20-2002-4b15-a6b6-f23874343914-run-httpd\") pod \"ceilometer-0\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " pod="openstack/ceilometer-0" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.077666 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " pod="openstack/ceilometer-0" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.077684 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-config-data\") pod \"ceilometer-0\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " pod="openstack/ceilometer-0" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.078612 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1df5e20-2002-4b15-a6b6-f23874343914-run-httpd\") pod \"ceilometer-0\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " pod="openstack/ceilometer-0" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.079440 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1df5e20-2002-4b15-a6b6-f23874343914-log-httpd\") pod \"ceilometer-0\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " pod="openstack/ceilometer-0" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.081898 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " pod="openstack/ceilometer-0" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.083895 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-config-data\") pod \"ceilometer-0\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " pod="openstack/ceilometer-0" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.085395 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-scripts\") pod \"ceilometer-0\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " pod="openstack/ceilometer-0" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.096704 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " pod="openstack/ceilometer-0" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.099904 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfzld\" (UniqueName: \"kubernetes.io/projected/e1df5e20-2002-4b15-a6b6-f23874343914-kube-api-access-bfzld\") pod \"ceilometer-0\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " pod="openstack/ceilometer-0" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.144772 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.605982 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.701648 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.734999 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1df5e20-2002-4b15-a6b6-f23874343914","Type":"ContainerStarted","Data":"327aa93255173331b392ba86f6b96ae35c2656bab40ba43f69c5ae070dd5c299"} Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.737955 4696 generic.go:334] "Generic (PLEG): container finished" podID="ae265c6e-25bb-421d-a717-78e42a3e38ca" containerID="c60e3de586d0628a13c96ac8a17d5a9162547031d23b7cd57a36307afbc59d61" exitCode=0 Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.738085 4696 generic.go:334] "Generic (PLEG): container finished" podID="ae265c6e-25bb-421d-a717-78e42a3e38ca" containerID="345e7ef002f347d2d9f10204eacfad5d403c1e573b451455e2599cb2a7224dd8" exitCode=143 Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.738248 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ae265c6e-25bb-421d-a717-78e42a3e38ca","Type":"ContainerDied","Data":"c60e3de586d0628a13c96ac8a17d5a9162547031d23b7cd57a36307afbc59d61"} Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.738372 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ae265c6e-25bb-421d-a717-78e42a3e38ca","Type":"ContainerDied","Data":"345e7ef002f347d2d9f10204eacfad5d403c1e573b451455e2599cb2a7224dd8"} Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.738489 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ae265c6e-25bb-421d-a717-78e42a3e38ca","Type":"ContainerDied","Data":"c1e1cf21e70d52111c086ff7b02fedf9b74d167eb95439180d4f7addbe589e22"} Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.738599 4696 scope.go:117] "RemoveContainer" containerID="c60e3de586d0628a13c96ac8a17d5a9162547031d23b7cd57a36307afbc59d61" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.738802 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.771214 4696 scope.go:117] "RemoveContainer" containerID="345e7ef002f347d2d9f10204eacfad5d403c1e573b451455e2599cb2a7224dd8" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.794682 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae265c6e-25bb-421d-a717-78e42a3e38ca-combined-ca-bundle\") pod \"ae265c6e-25bb-421d-a717-78e42a3e38ca\" (UID: \"ae265c6e-25bb-421d-a717-78e42a3e38ca\") " Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.794915 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae265c6e-25bb-421d-a717-78e42a3e38ca-logs\") pod \"ae265c6e-25bb-421d-a717-78e42a3e38ca\" (UID: \"ae265c6e-25bb-421d-a717-78e42a3e38ca\") " Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.794957 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jk4wl\" (UniqueName: \"kubernetes.io/projected/ae265c6e-25bb-421d-a717-78e42a3e38ca-kube-api-access-jk4wl\") pod \"ae265c6e-25bb-421d-a717-78e42a3e38ca\" (UID: \"ae265c6e-25bb-421d-a717-78e42a3e38ca\") " Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.794979 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae265c6e-25bb-421d-a717-78e42a3e38ca-config-data\") pod \"ae265c6e-25bb-421d-a717-78e42a3e38ca\" (UID: \"ae265c6e-25bb-421d-a717-78e42a3e38ca\") " Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.795734 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae265c6e-25bb-421d-a717-78e42a3e38ca-logs" (OuterVolumeSpecName: "logs") pod "ae265c6e-25bb-421d-a717-78e42a3e38ca" (UID: "ae265c6e-25bb-421d-a717-78e42a3e38ca"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.801759 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae265c6e-25bb-421d-a717-78e42a3e38ca-kube-api-access-jk4wl" (OuterVolumeSpecName: "kube-api-access-jk4wl") pod "ae265c6e-25bb-421d-a717-78e42a3e38ca" (UID: "ae265c6e-25bb-421d-a717-78e42a3e38ca"). InnerVolumeSpecName "kube-api-access-jk4wl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.806175 4696 scope.go:117] "RemoveContainer" containerID="c60e3de586d0628a13c96ac8a17d5a9162547031d23b7cd57a36307afbc59d61" Mar 21 08:49:33 crc kubenswrapper[4696]: E0321 08:49:33.806694 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c60e3de586d0628a13c96ac8a17d5a9162547031d23b7cd57a36307afbc59d61\": container with ID starting with c60e3de586d0628a13c96ac8a17d5a9162547031d23b7cd57a36307afbc59d61 not found: ID does not exist" containerID="c60e3de586d0628a13c96ac8a17d5a9162547031d23b7cd57a36307afbc59d61" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.806734 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c60e3de586d0628a13c96ac8a17d5a9162547031d23b7cd57a36307afbc59d61"} err="failed to get container status \"c60e3de586d0628a13c96ac8a17d5a9162547031d23b7cd57a36307afbc59d61\": rpc error: code = NotFound desc = could not find container \"c60e3de586d0628a13c96ac8a17d5a9162547031d23b7cd57a36307afbc59d61\": container with ID starting with c60e3de586d0628a13c96ac8a17d5a9162547031d23b7cd57a36307afbc59d61 not found: ID does not exist" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.806764 4696 scope.go:117] "RemoveContainer" containerID="345e7ef002f347d2d9f10204eacfad5d403c1e573b451455e2599cb2a7224dd8" Mar 21 08:49:33 crc kubenswrapper[4696]: E0321 08:49:33.808323 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"345e7ef002f347d2d9f10204eacfad5d403c1e573b451455e2599cb2a7224dd8\": container with ID starting with 345e7ef002f347d2d9f10204eacfad5d403c1e573b451455e2599cb2a7224dd8 not found: ID does not exist" containerID="345e7ef002f347d2d9f10204eacfad5d403c1e573b451455e2599cb2a7224dd8" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.808358 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"345e7ef002f347d2d9f10204eacfad5d403c1e573b451455e2599cb2a7224dd8"} err="failed to get container status \"345e7ef002f347d2d9f10204eacfad5d403c1e573b451455e2599cb2a7224dd8\": rpc error: code = NotFound desc = could not find container \"345e7ef002f347d2d9f10204eacfad5d403c1e573b451455e2599cb2a7224dd8\": container with ID starting with 345e7ef002f347d2d9f10204eacfad5d403c1e573b451455e2599cb2a7224dd8 not found: ID does not exist" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.808381 4696 scope.go:117] "RemoveContainer" containerID="c60e3de586d0628a13c96ac8a17d5a9162547031d23b7cd57a36307afbc59d61" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.810044 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c60e3de586d0628a13c96ac8a17d5a9162547031d23b7cd57a36307afbc59d61"} err="failed to get container status \"c60e3de586d0628a13c96ac8a17d5a9162547031d23b7cd57a36307afbc59d61\": rpc error: code = NotFound desc = could not find container \"c60e3de586d0628a13c96ac8a17d5a9162547031d23b7cd57a36307afbc59d61\": container with ID starting with c60e3de586d0628a13c96ac8a17d5a9162547031d23b7cd57a36307afbc59d61 not found: ID does not exist" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.810075 4696 scope.go:117] "RemoveContainer" containerID="345e7ef002f347d2d9f10204eacfad5d403c1e573b451455e2599cb2a7224dd8" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.812175 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"345e7ef002f347d2d9f10204eacfad5d403c1e573b451455e2599cb2a7224dd8"} err="failed to get container status \"345e7ef002f347d2d9f10204eacfad5d403c1e573b451455e2599cb2a7224dd8\": rpc error: code = NotFound desc = could not find container \"345e7ef002f347d2d9f10204eacfad5d403c1e573b451455e2599cb2a7224dd8\": container with ID starting with 345e7ef002f347d2d9f10204eacfad5d403c1e573b451455e2599cb2a7224dd8 not found: ID does not exist" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.835517 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae265c6e-25bb-421d-a717-78e42a3e38ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae265c6e-25bb-421d-a717-78e42a3e38ca" (UID: "ae265c6e-25bb-421d-a717-78e42a3e38ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.847347 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae265c6e-25bb-421d-a717-78e42a3e38ca-config-data" (OuterVolumeSpecName: "config-data") pod "ae265c6e-25bb-421d-a717-78e42a3e38ca" (UID: "ae265c6e-25bb-421d-a717-78e42a3e38ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.897730 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae265c6e-25bb-421d-a717-78e42a3e38ca-logs\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.897763 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jk4wl\" (UniqueName: \"kubernetes.io/projected/ae265c6e-25bb-421d-a717-78e42a3e38ca-kube-api-access-jk4wl\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.897775 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae265c6e-25bb-421d-a717-78e42a3e38ca-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:33 crc kubenswrapper[4696]: I0321 08:49:33.897784 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae265c6e-25bb-421d-a717-78e42a3e38ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.088593 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.098529 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.115884 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 21 08:49:34 crc kubenswrapper[4696]: E0321 08:49:34.116496 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae265c6e-25bb-421d-a717-78e42a3e38ca" containerName="nova-metadata-metadata" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.116512 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae265c6e-25bb-421d-a717-78e42a3e38ca" containerName="nova-metadata-metadata" Mar 21 08:49:34 crc kubenswrapper[4696]: E0321 08:49:34.116527 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae265c6e-25bb-421d-a717-78e42a3e38ca" containerName="nova-metadata-log" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.116534 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae265c6e-25bb-421d-a717-78e42a3e38ca" containerName="nova-metadata-log" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.116776 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae265c6e-25bb-421d-a717-78e42a3e38ca" containerName="nova-metadata-metadata" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.116799 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae265c6e-25bb-421d-a717-78e42a3e38ca" containerName="nova-metadata-log" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.119614 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.124533 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.124681 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.149400 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.309151 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-logs\") pod \"nova-metadata-0\" (UID: \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\") " pod="openstack/nova-metadata-0" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.309220 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-config-data\") pod \"nova-metadata-0\" (UID: \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\") " pod="openstack/nova-metadata-0" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.309300 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzhq4\" (UniqueName: \"kubernetes.io/projected/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-kube-api-access-fzhq4\") pod \"nova-metadata-0\" (UID: \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\") " pod="openstack/nova-metadata-0" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.309327 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\") " pod="openstack/nova-metadata-0" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.309391 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\") " pod="openstack/nova-metadata-0" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.410748 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\") " pod="openstack/nova-metadata-0" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.410865 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-logs\") pod \"nova-metadata-0\" (UID: \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\") " pod="openstack/nova-metadata-0" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.410910 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-config-data\") pod \"nova-metadata-0\" (UID: \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\") " pod="openstack/nova-metadata-0" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.411008 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzhq4\" (UniqueName: \"kubernetes.io/projected/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-kube-api-access-fzhq4\") pod \"nova-metadata-0\" (UID: \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\") " pod="openstack/nova-metadata-0" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.411038 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\") " pod="openstack/nova-metadata-0" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.411473 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-logs\") pod \"nova-metadata-0\" (UID: \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\") " pod="openstack/nova-metadata-0" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.414924 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\") " pod="openstack/nova-metadata-0" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.414992 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-config-data\") pod \"nova-metadata-0\" (UID: \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\") " pod="openstack/nova-metadata-0" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.429290 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\") " pod="openstack/nova-metadata-0" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.430099 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzhq4\" (UniqueName: \"kubernetes.io/projected/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-kube-api-access-fzhq4\") pod \"nova-metadata-0\" (UID: \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\") " pod="openstack/nova-metadata-0" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.545414 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f9d152a-965e-417f-9d2a-5ee0d411cf12" path="/var/lib/kubelet/pods/8f9d152a-965e-417f-9d2a-5ee0d411cf12/volumes" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.546477 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae265c6e-25bb-421d-a717-78e42a3e38ca" path="/var/lib/kubelet/pods/ae265c6e-25bb-421d-a717-78e42a3e38ca/volumes" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.685691 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.689927 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.690052 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.835192 4696 generic.go:334] "Generic (PLEG): container finished" podID="29e6fe18-05b8-430f-b3fe-96ca1cf04a94" containerID="32c9d7bf028548a11c13896a74b2d7e729721772af461048e7575f7de03057ed" exitCode=0 Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.835530 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-d4v88" event={"ID":"29e6fe18-05b8-430f-b3fe-96ca1cf04a94","Type":"ContainerDied","Data":"32c9d7bf028548a11c13896a74b2d7e729721772af461048e7575f7de03057ed"} Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.840636 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.847517 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1df5e20-2002-4b15-a6b6-f23874343914","Type":"ContainerStarted","Data":"bece433f574d7f51b9470e8875dfab386d4755bbb68df8962d52b36c582ba420"} Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.975485 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 21 08:49:34 crc kubenswrapper[4696]: I0321 08:49:34.976357 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 21 08:49:35 crc kubenswrapper[4696]: I0321 08:49:35.015428 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:49:35 crc kubenswrapper[4696]: I0321 08:49:35.038960 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 21 08:49:35 crc kubenswrapper[4696]: I0321 08:49:35.157615 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67bdc55879-kc4xr"] Mar 21 08:49:35 crc kubenswrapper[4696]: I0321 08:49:35.162489 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" podUID="2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1" containerName="dnsmasq-dns" containerID="cri-o://ade47d5d8b5fc01f442812f9ccf8b8af5c438979e8d544094bdc36466086ef07" gracePeriod=10 Mar 21 08:49:35 crc kubenswrapper[4696]: I0321 08:49:35.238708 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 08:49:35 crc kubenswrapper[4696]: W0321 08:49:35.252032 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab19b72e_ead0_4baa_ba2a_badfae78ffa0.slice/crio-625a5a7359df21935e8f8d6d7d59609fa048e73b5e5dc2871b2f39cb3339b914 WatchSource:0}: Error finding container 625a5a7359df21935e8f8d6d7d59609fa048e73b5e5dc2871b2f39cb3339b914: Status 404 returned error can't find the container with id 625a5a7359df21935e8f8d6d7d59609fa048e73b5e5dc2871b2f39cb3339b914 Mar 21 08:49:35 crc kubenswrapper[4696]: I0321 08:49:35.774330 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.218:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 21 08:49:35 crc kubenswrapper[4696]: I0321 08:49:35.774348 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.218:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 21 08:49:35 crc kubenswrapper[4696]: I0321 08:49:35.871870 4696 generic.go:334] "Generic (PLEG): container finished" podID="2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1" containerID="ade47d5d8b5fc01f442812f9ccf8b8af5c438979e8d544094bdc36466086ef07" exitCode=0 Mar 21 08:49:35 crc kubenswrapper[4696]: I0321 08:49:35.871963 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" event={"ID":"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1","Type":"ContainerDied","Data":"ade47d5d8b5fc01f442812f9ccf8b8af5c438979e8d544094bdc36466086ef07"} Mar 21 08:49:35 crc kubenswrapper[4696]: I0321 08:49:35.879245 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ab19b72e-ead0-4baa-ba2a-badfae78ffa0","Type":"ContainerStarted","Data":"9f40ef8be4528f0dd40034174d9f596fc019975a89b1b58e178f8d5b86bb7fdc"} Mar 21 08:49:35 crc kubenswrapper[4696]: I0321 08:49:35.879299 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ab19b72e-ead0-4baa-ba2a-badfae78ffa0","Type":"ContainerStarted","Data":"625a5a7359df21935e8f8d6d7d59609fa048e73b5e5dc2871b2f39cb3339b914"} Mar 21 08:49:35 crc kubenswrapper[4696]: I0321 08:49:35.882595 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1df5e20-2002-4b15-a6b6-f23874343914","Type":"ContainerStarted","Data":"7eb1093e5d953b48d0bcdd3d193e4c4a05d3dd7fc4246704a7cee0195e1cb843"} Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.036374 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.318223 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.378336 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-dns-svc\") pod \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.378520 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-ovsdbserver-sb\") pod \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.378555 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-dns-swift-storage-0\") pod \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.378620 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gw2p\" (UniqueName: \"kubernetes.io/projected/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-kube-api-access-7gw2p\") pod \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.378698 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-config\") pod \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.378715 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-ovsdbserver-nb\") pod \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\" (UID: \"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1\") " Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.411207 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-kube-api-access-7gw2p" (OuterVolumeSpecName: "kube-api-access-7gw2p") pod "2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1" (UID: "2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1"). InnerVolumeSpecName "kube-api-access-7gw2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.486222 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gw2p\" (UniqueName: \"kubernetes.io/projected/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-kube-api-access-7gw2p\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.666423 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1" (UID: "2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.690252 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.731295 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1" (UID: "2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.733446 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1" (UID: "2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.739036 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1" (UID: "2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.742416 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-d4v88" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.749461 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-config" (OuterVolumeSpecName: "config") pod "2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1" (UID: "2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.793098 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.793126 4696 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.793138 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.793146 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.896909 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zncj4\" (UniqueName: \"kubernetes.io/projected/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-kube-api-access-zncj4\") pod \"29e6fe18-05b8-430f-b3fe-96ca1cf04a94\" (UID: \"29e6fe18-05b8-430f-b3fe-96ca1cf04a94\") " Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.897044 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-config-data\") pod \"29e6fe18-05b8-430f-b3fe-96ca1cf04a94\" (UID: \"29e6fe18-05b8-430f-b3fe-96ca1cf04a94\") " Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.897098 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-combined-ca-bundle\") pod \"29e6fe18-05b8-430f-b3fe-96ca1cf04a94\" (UID: \"29e6fe18-05b8-430f-b3fe-96ca1cf04a94\") " Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.897123 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-scripts\") pod \"29e6fe18-05b8-430f-b3fe-96ca1cf04a94\" (UID: \"29e6fe18-05b8-430f-b3fe-96ca1cf04a94\") " Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.912107 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-kube-api-access-zncj4" (OuterVolumeSpecName: "kube-api-access-zncj4") pod "29e6fe18-05b8-430f-b3fe-96ca1cf04a94" (UID: "29e6fe18-05b8-430f-b3fe-96ca1cf04a94"). InnerVolumeSpecName "kube-api-access-zncj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.912245 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-scripts" (OuterVolumeSpecName: "scripts") pod "29e6fe18-05b8-430f-b3fe-96ca1cf04a94" (UID: "29e6fe18-05b8-430f-b3fe-96ca1cf04a94"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.912702 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ab19b72e-ead0-4baa-ba2a-badfae78ffa0","Type":"ContainerStarted","Data":"7f5a53b6a94c2e78a2c9bf427a0fa63cba9763534bcd0f90e2556589532c5c60"} Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.927252 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-d4v88" event={"ID":"29e6fe18-05b8-430f-b3fe-96ca1cf04a94","Type":"ContainerDied","Data":"27f55643d95b74efe15532bc302311a0a1e896bee3df443a4421d813c5214d51"} Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.927305 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27f55643d95b74efe15532bc302311a0a1e896bee3df443a4421d813c5214d51" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.927388 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-d4v88" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.939173 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1df5e20-2002-4b15-a6b6-f23874343914","Type":"ContainerStarted","Data":"7fbc416c207de813688566084124df0e65f347c6de903ae1dcdf3d7a33fdb3e4"} Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.942261 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "29e6fe18-05b8-430f-b3fe-96ca1cf04a94" (UID: "29e6fe18-05b8-430f-b3fe-96ca1cf04a94"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.949769 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.949751364 podStartE2EDuration="2.949751364s" podCreationTimestamp="2026-03-21 08:49:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:49:36.934944364 +0000 UTC m=+1311.055825087" watchObservedRunningTime="2026-03-21 08:49:36.949751364 +0000 UTC m=+1311.070632077" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.954838 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.959852 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" event={"ID":"2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1","Type":"ContainerDied","Data":"257c3c64522b08c742fb41cd2edcc8e6bd95b7b76a449c671fce1f6ec1963b0b"} Mar 21 08:49:36 crc kubenswrapper[4696]: I0321 08:49:36.959941 4696 scope.go:117] "RemoveContainer" containerID="ade47d5d8b5fc01f442812f9ccf8b8af5c438979e8d544094bdc36466086ef07" Mar 21 08:49:37 crc kubenswrapper[4696]: I0321 08:49:36.997544 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67bdc55879-kc4xr"] Mar 21 08:49:37 crc kubenswrapper[4696]: I0321 08:49:37.004970 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zncj4\" (UniqueName: \"kubernetes.io/projected/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-kube-api-access-zncj4\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:37 crc kubenswrapper[4696]: I0321 08:49:37.005008 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:37 crc kubenswrapper[4696]: I0321 08:49:37.005017 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:37 crc kubenswrapper[4696]: I0321 08:49:37.008027 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-config-data" (OuterVolumeSpecName: "config-data") pod "29e6fe18-05b8-430f-b3fe-96ca1cf04a94" (UID: "29e6fe18-05b8-430f-b3fe-96ca1cf04a94"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:37 crc kubenswrapper[4696]: I0321 08:49:37.044003 4696 scope.go:117] "RemoveContainer" containerID="02b2c669c853f02bcbddcb5cce028bbad7b960d0821dd2cd4d12972bd0670824" Mar 21 08:49:37 crc kubenswrapper[4696]: I0321 08:49:37.051286 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 21 08:49:37 crc kubenswrapper[4696]: I0321 08:49:37.051786 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1" containerName="nova-api-log" containerID="cri-o://a5078c0880aedda106502b88c3ed962c04f889b9117d6af7d22c03cccf1fc203" gracePeriod=30 Mar 21 08:49:37 crc kubenswrapper[4696]: I0321 08:49:37.052548 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1" containerName="nova-api-api" containerID="cri-o://73da593316c9714471379b39652eebf6cf072a108dd7bbee42f1ef069967be5f" gracePeriod=30 Mar 21 08:49:37 crc kubenswrapper[4696]: I0321 08:49:37.059827 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67bdc55879-kc4xr"] Mar 21 08:49:37 crc kubenswrapper[4696]: I0321 08:49:37.110761 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29e6fe18-05b8-430f-b3fe-96ca1cf04a94-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:37 crc kubenswrapper[4696]: I0321 08:49:37.138155 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 08:49:37 crc kubenswrapper[4696]: E0321 08:49:37.736162 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4856e60e_c40e_4b95_bc64_2f33b8b79e4b.slice/crio-conmon-4be79300257c4c766e3f6f08ee86a69724e964c534ae69af90e17eeac61e8aa6.scope\": RecentStats: unable to find data in memory cache]" Mar 21 08:49:37 crc kubenswrapper[4696]: I0321 08:49:37.951200 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 08:49:37 crc kubenswrapper[4696]: I0321 08:49:37.967168 4696 generic.go:334] "Generic (PLEG): container finished" podID="4856e60e-c40e-4b95-bc64-2f33b8b79e4b" containerID="4be79300257c4c766e3f6f08ee86a69724e964c534ae69af90e17eeac61e8aa6" exitCode=0 Mar 21 08:49:37 crc kubenswrapper[4696]: I0321 08:49:37.967250 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fjcrs" event={"ID":"4856e60e-c40e-4b95-bc64-2f33b8b79e4b","Type":"ContainerDied","Data":"4be79300257c4c766e3f6f08ee86a69724e964c534ae69af90e17eeac61e8aa6"} Mar 21 08:49:37 crc kubenswrapper[4696]: I0321 08:49:37.969031 4696 generic.go:334] "Generic (PLEG): container finished" podID="34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1" containerID="a5078c0880aedda106502b88c3ed962c04f889b9117d6af7d22c03cccf1fc203" exitCode=143 Mar 21 08:49:37 crc kubenswrapper[4696]: I0321 08:49:37.969099 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1","Type":"ContainerDied","Data":"a5078c0880aedda106502b88c3ed962c04f889b9117d6af7d22c03cccf1fc203"} Mar 21 08:49:37 crc kubenswrapper[4696]: I0321 08:49:37.969286 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="7a1d264d-7e6b-4910-9656-61ffefe53f8d" containerName="nova-scheduler-scheduler" containerID="cri-o://92ad6b8cfc9519d7f9031eec9efa261d2fe191665ac15c9fca809d3ecc7e6279" gracePeriod=30 Mar 21 08:49:38 crc kubenswrapper[4696]: I0321 08:49:38.545467 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1" path="/var/lib/kubelet/pods/2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1/volumes" Mar 21 08:49:38 crc kubenswrapper[4696]: I0321 08:49:38.981839 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ab19b72e-ead0-4baa-ba2a-badfae78ffa0" containerName="nova-metadata-log" containerID="cri-o://9f40ef8be4528f0dd40034174d9f596fc019975a89b1b58e178f8d5b86bb7fdc" gracePeriod=30 Mar 21 08:49:38 crc kubenswrapper[4696]: I0321 08:49:38.983317 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1df5e20-2002-4b15-a6b6-f23874343914","Type":"ContainerStarted","Data":"52712668b4c83dc1d0113247da7073898d3f4a1431b21f117f82c1a4d0118d02"} Mar 21 08:49:38 crc kubenswrapper[4696]: I0321 08:49:38.983365 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 21 08:49:38 crc kubenswrapper[4696]: I0321 08:49:38.984105 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ab19b72e-ead0-4baa-ba2a-badfae78ffa0" containerName="nova-metadata-metadata" containerID="cri-o://7f5a53b6a94c2e78a2c9bf427a0fa63cba9763534bcd0f90e2556589532c5c60" gracePeriod=30 Mar 21 08:49:39 crc kubenswrapper[4696]: I0321 08:49:39.025438 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.497266201 podStartE2EDuration="7.025417083s" podCreationTimestamp="2026-03-21 08:49:32 +0000 UTC" firstStartedPulling="2026-03-21 08:49:33.704501928 +0000 UTC m=+1307.825382641" lastFinishedPulling="2026-03-21 08:49:38.23265281 +0000 UTC m=+1312.353533523" observedRunningTime="2026-03-21 08:49:39.015810696 +0000 UTC m=+1313.136691429" watchObservedRunningTime="2026-03-21 08:49:39.025417083 +0000 UTC m=+1313.146297796" Mar 21 08:49:39 crc kubenswrapper[4696]: I0321 08:49:39.801540 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fjcrs" Mar 21 08:49:39 crc kubenswrapper[4696]: I0321 08:49:39.869145 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmvwv\" (UniqueName: \"kubernetes.io/projected/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-kube-api-access-xmvwv\") pod \"4856e60e-c40e-4b95-bc64-2f33b8b79e4b\" (UID: \"4856e60e-c40e-4b95-bc64-2f33b8b79e4b\") " Mar 21 08:49:39 crc kubenswrapper[4696]: I0321 08:49:39.869192 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-scripts\") pod \"4856e60e-c40e-4b95-bc64-2f33b8b79e4b\" (UID: \"4856e60e-c40e-4b95-bc64-2f33b8b79e4b\") " Mar 21 08:49:39 crc kubenswrapper[4696]: I0321 08:49:39.869233 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-config-data\") pod \"4856e60e-c40e-4b95-bc64-2f33b8b79e4b\" (UID: \"4856e60e-c40e-4b95-bc64-2f33b8b79e4b\") " Mar 21 08:49:39 crc kubenswrapper[4696]: I0321 08:49:39.869300 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-combined-ca-bundle\") pod \"4856e60e-c40e-4b95-bc64-2f33b8b79e4b\" (UID: \"4856e60e-c40e-4b95-bc64-2f33b8b79e4b\") " Mar 21 08:49:39 crc kubenswrapper[4696]: I0321 08:49:39.880949 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-scripts" (OuterVolumeSpecName: "scripts") pod "4856e60e-c40e-4b95-bc64-2f33b8b79e4b" (UID: "4856e60e-c40e-4b95-bc64-2f33b8b79e4b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:39 crc kubenswrapper[4696]: I0321 08:49:39.882211 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-kube-api-access-xmvwv" (OuterVolumeSpecName: "kube-api-access-xmvwv") pod "4856e60e-c40e-4b95-bc64-2f33b8b79e4b" (UID: "4856e60e-c40e-4b95-bc64-2f33b8b79e4b"). InnerVolumeSpecName "kube-api-access-xmvwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:49:39 crc kubenswrapper[4696]: I0321 08:49:39.920648 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-config-data" (OuterVolumeSpecName: "config-data") pod "4856e60e-c40e-4b95-bc64-2f33b8b79e4b" (UID: "4856e60e-c40e-4b95-bc64-2f33b8b79e4b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:39 crc kubenswrapper[4696]: I0321 08:49:39.924710 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4856e60e-c40e-4b95-bc64-2f33b8b79e4b" (UID: "4856e60e-c40e-4b95-bc64-2f33b8b79e4b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:39 crc kubenswrapper[4696]: I0321 08:49:39.977142 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmvwv\" (UniqueName: \"kubernetes.io/projected/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-kube-api-access-xmvwv\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:39 crc kubenswrapper[4696]: I0321 08:49:39.977174 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:39 crc kubenswrapper[4696]: I0321 08:49:39.977184 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:39 crc kubenswrapper[4696]: I0321 08:49:39.977192 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4856e60e-c40e-4b95-bc64-2f33b8b79e4b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:39 crc kubenswrapper[4696]: E0321 08:49:39.978574 4696 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="92ad6b8cfc9519d7f9031eec9efa261d2fe191665ac15c9fca809d3ecc7e6279" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 21 08:49:39 crc kubenswrapper[4696]: E0321 08:49:39.980428 4696 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="92ad6b8cfc9519d7f9031eec9efa261d2fe191665ac15c9fca809d3ecc7e6279" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 21 08:49:39 crc kubenswrapper[4696]: E0321 08:49:39.982304 4696 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="92ad6b8cfc9519d7f9031eec9efa261d2fe191665ac15c9fca809d3ecc7e6279" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 21 08:49:39 crc kubenswrapper[4696]: E0321 08:49:39.982340 4696 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="7a1d264d-7e6b-4910-9656-61ffefe53f8d" containerName="nova-scheduler-scheduler" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.026429 4696 generic.go:334] "Generic (PLEG): container finished" podID="ab19b72e-ead0-4baa-ba2a-badfae78ffa0" containerID="7f5a53b6a94c2e78a2c9bf427a0fa63cba9763534bcd0f90e2556589532c5c60" exitCode=0 Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.026457 4696 generic.go:334] "Generic (PLEG): container finished" podID="ab19b72e-ead0-4baa-ba2a-badfae78ffa0" containerID="9f40ef8be4528f0dd40034174d9f596fc019975a89b1b58e178f8d5b86bb7fdc" exitCode=143 Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.026495 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ab19b72e-ead0-4baa-ba2a-badfae78ffa0","Type":"ContainerDied","Data":"7f5a53b6a94c2e78a2c9bf427a0fa63cba9763534bcd0f90e2556589532c5c60"} Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.026528 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ab19b72e-ead0-4baa-ba2a-badfae78ffa0","Type":"ContainerDied","Data":"9f40ef8be4528f0dd40034174d9f596fc019975a89b1b58e178f8d5b86bb7fdc"} Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.029283 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fjcrs" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.029937 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fjcrs" event={"ID":"4856e60e-c40e-4b95-bc64-2f33b8b79e4b","Type":"ContainerDied","Data":"09b52a120a53a4ed65c3df82da374baa2820634eaaff50e517bf1a3719cd1381"} Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.029994 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09b52a120a53a4ed65c3df82da374baa2820634eaaff50e517bf1a3719cd1381" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.098703 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 21 08:49:40 crc kubenswrapper[4696]: E0321 08:49:40.099348 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1" containerName="dnsmasq-dns" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.099368 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1" containerName="dnsmasq-dns" Mar 21 08:49:40 crc kubenswrapper[4696]: E0321 08:49:40.099394 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4856e60e-c40e-4b95-bc64-2f33b8b79e4b" containerName="nova-cell1-conductor-db-sync" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.099401 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4856e60e-c40e-4b95-bc64-2f33b8b79e4b" containerName="nova-cell1-conductor-db-sync" Mar 21 08:49:40 crc kubenswrapper[4696]: E0321 08:49:40.099413 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1" containerName="init" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.099419 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1" containerName="init" Mar 21 08:49:40 crc kubenswrapper[4696]: E0321 08:49:40.099466 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29e6fe18-05b8-430f-b3fe-96ca1cf04a94" containerName="nova-manage" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.099473 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="29e6fe18-05b8-430f-b3fe-96ca1cf04a94" containerName="nova-manage" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.099804 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="29e6fe18-05b8-430f-b3fe-96ca1cf04a94" containerName="nova-manage" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.099843 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="4856e60e-c40e-4b95-bc64-2f33b8b79e4b" containerName="nova-cell1-conductor-db-sync" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.099865 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1" containerName="dnsmasq-dns" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.100790 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.103144 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.117498 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.201452 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267\") " pod="openstack/nova-cell1-conductor-0" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.202317 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267\") " pod="openstack/nova-cell1-conductor-0" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.202403 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghhtj\" (UniqueName: \"kubernetes.io/projected/58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267-kube-api-access-ghhtj\") pod \"nova-cell1-conductor-0\" (UID: \"58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267\") " pod="openstack/nova-cell1-conductor-0" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.305882 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267\") " pod="openstack/nova-cell1-conductor-0" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.306119 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267\") " pod="openstack/nova-cell1-conductor-0" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.306516 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghhtj\" (UniqueName: \"kubernetes.io/projected/58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267-kube-api-access-ghhtj\") pod \"nova-cell1-conductor-0\" (UID: \"58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267\") " pod="openstack/nova-cell1-conductor-0" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.311273 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267\") " pod="openstack/nova-cell1-conductor-0" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.311873 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267\") " pod="openstack/nova-cell1-conductor-0" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.329285 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghhtj\" (UniqueName: \"kubernetes.io/projected/58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267-kube-api-access-ghhtj\") pod \"nova-cell1-conductor-0\" (UID: \"58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267\") " pod="openstack/nova-cell1-conductor-0" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.433549 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.499098 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-67bdc55879-kc4xr" podUID="2fda15a2-d7ce-4c10-8dc8-961fc12cc2a1" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.194:5353: i/o timeout" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.712079 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.817482 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-config-data\") pod \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\" (UID: \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\") " Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.817802 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-combined-ca-bundle\") pod \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\" (UID: \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\") " Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.818031 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzhq4\" (UniqueName: \"kubernetes.io/projected/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-kube-api-access-fzhq4\") pod \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\" (UID: \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\") " Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.818144 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-logs\") pod \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\" (UID: \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\") " Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.818339 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-nova-metadata-tls-certs\") pod \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\" (UID: \"ab19b72e-ead0-4baa-ba2a-badfae78ffa0\") " Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.818508 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-logs" (OuterVolumeSpecName: "logs") pod "ab19b72e-ead0-4baa-ba2a-badfae78ffa0" (UID: "ab19b72e-ead0-4baa-ba2a-badfae78ffa0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.819071 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-logs\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.823552 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-kube-api-access-fzhq4" (OuterVolumeSpecName: "kube-api-access-fzhq4") pod "ab19b72e-ead0-4baa-ba2a-badfae78ffa0" (UID: "ab19b72e-ead0-4baa-ba2a-badfae78ffa0"). InnerVolumeSpecName "kube-api-access-fzhq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.879831 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-config-data" (OuterVolumeSpecName: "config-data") pod "ab19b72e-ead0-4baa-ba2a-badfae78ffa0" (UID: "ab19b72e-ead0-4baa-ba2a-badfae78ffa0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.906120 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "ab19b72e-ead0-4baa-ba2a-badfae78ffa0" (UID: "ab19b72e-ead0-4baa-ba2a-badfae78ffa0"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.920836 4696 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.920871 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.920883 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzhq4\" (UniqueName: \"kubernetes.io/projected/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-kube-api-access-fzhq4\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:40 crc kubenswrapper[4696]: I0321 08:49:40.939942 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab19b72e-ead0-4baa-ba2a-badfae78ffa0" (UID: "ab19b72e-ead0-4baa-ba2a-badfae78ffa0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.019849 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.023582 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab19b72e-ead0-4baa-ba2a-badfae78ffa0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.050438 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.050450 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ab19b72e-ead0-4baa-ba2a-badfae78ffa0","Type":"ContainerDied","Data":"625a5a7359df21935e8f8d6d7d59609fa048e73b5e5dc2871b2f39cb3339b914"} Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.050501 4696 scope.go:117] "RemoveContainer" containerID="7f5a53b6a94c2e78a2c9bf427a0fa63cba9763534bcd0f90e2556589532c5c60" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.058473 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267","Type":"ContainerStarted","Data":"85a8ef3064c4b451c6d8e022e4df64f76b4cbbab7c951756edcb7838465e40f1"} Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.062710 4696 generic.go:334] "Generic (PLEG): container finished" podID="7a1d264d-7e6b-4910-9656-61ffefe53f8d" containerID="92ad6b8cfc9519d7f9031eec9efa261d2fe191665ac15c9fca809d3ecc7e6279" exitCode=0 Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.062741 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7a1d264d-7e6b-4910-9656-61ffefe53f8d","Type":"ContainerDied","Data":"92ad6b8cfc9519d7f9031eec9efa261d2fe191665ac15c9fca809d3ecc7e6279"} Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.108293 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.128148 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.139607 4696 scope.go:117] "RemoveContainer" containerID="9f40ef8be4528f0dd40034174d9f596fc019975a89b1b58e178f8d5b86bb7fdc" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.153987 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 21 08:49:41 crc kubenswrapper[4696]: E0321 08:49:41.154500 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab19b72e-ead0-4baa-ba2a-badfae78ffa0" containerName="nova-metadata-metadata" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.154519 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab19b72e-ead0-4baa-ba2a-badfae78ffa0" containerName="nova-metadata-metadata" Mar 21 08:49:41 crc kubenswrapper[4696]: E0321 08:49:41.154539 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab19b72e-ead0-4baa-ba2a-badfae78ffa0" containerName="nova-metadata-log" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.154547 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab19b72e-ead0-4baa-ba2a-badfae78ffa0" containerName="nova-metadata-log" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.154766 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab19b72e-ead0-4baa-ba2a-badfae78ffa0" containerName="nova-metadata-log" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.154792 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab19b72e-ead0-4baa-ba2a-badfae78ffa0" containerName="nova-metadata-metadata" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.157924 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.161319 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.161560 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.191917 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.231536 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\") " pod="openstack/nova-metadata-0" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.231867 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-config-data\") pod \"nova-metadata-0\" (UID: \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\") " pod="openstack/nova-metadata-0" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.231960 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-logs\") pod \"nova-metadata-0\" (UID: \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\") " pod="openstack/nova-metadata-0" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.232048 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdxq6\" (UniqueName: \"kubernetes.io/projected/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-kube-api-access-sdxq6\") pod \"nova-metadata-0\" (UID: \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\") " pod="openstack/nova-metadata-0" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.232129 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\") " pod="openstack/nova-metadata-0" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.307183 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.334053 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-config-data\") pod \"nova-metadata-0\" (UID: \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\") " pod="openstack/nova-metadata-0" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.334679 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-logs\") pod \"nova-metadata-0\" (UID: \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\") " pod="openstack/nova-metadata-0" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.334774 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdxq6\" (UniqueName: \"kubernetes.io/projected/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-kube-api-access-sdxq6\") pod \"nova-metadata-0\" (UID: \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\") " pod="openstack/nova-metadata-0" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.334881 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\") " pod="openstack/nova-metadata-0" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.335042 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\") " pod="openstack/nova-metadata-0" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.335087 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-logs\") pod \"nova-metadata-0\" (UID: \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\") " pod="openstack/nova-metadata-0" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.339202 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-config-data\") pod \"nova-metadata-0\" (UID: \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\") " pod="openstack/nova-metadata-0" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.339997 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\") " pod="openstack/nova-metadata-0" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.340255 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\") " pod="openstack/nova-metadata-0" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.352306 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdxq6\" (UniqueName: \"kubernetes.io/projected/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-kube-api-access-sdxq6\") pod \"nova-metadata-0\" (UID: \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\") " pod="openstack/nova-metadata-0" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.436439 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a1d264d-7e6b-4910-9656-61ffefe53f8d-config-data\") pod \"7a1d264d-7e6b-4910-9656-61ffefe53f8d\" (UID: \"7a1d264d-7e6b-4910-9656-61ffefe53f8d\") " Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.436746 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4s5z\" (UniqueName: \"kubernetes.io/projected/7a1d264d-7e6b-4910-9656-61ffefe53f8d-kube-api-access-r4s5z\") pod \"7a1d264d-7e6b-4910-9656-61ffefe53f8d\" (UID: \"7a1d264d-7e6b-4910-9656-61ffefe53f8d\") " Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.437008 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a1d264d-7e6b-4910-9656-61ffefe53f8d-combined-ca-bundle\") pod \"7a1d264d-7e6b-4910-9656-61ffefe53f8d\" (UID: \"7a1d264d-7e6b-4910-9656-61ffefe53f8d\") " Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.444960 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a1d264d-7e6b-4910-9656-61ffefe53f8d-kube-api-access-r4s5z" (OuterVolumeSpecName: "kube-api-access-r4s5z") pod "7a1d264d-7e6b-4910-9656-61ffefe53f8d" (UID: "7a1d264d-7e6b-4910-9656-61ffefe53f8d"). InnerVolumeSpecName "kube-api-access-r4s5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.476088 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a1d264d-7e6b-4910-9656-61ffefe53f8d-config-data" (OuterVolumeSpecName: "config-data") pod "7a1d264d-7e6b-4910-9656-61ffefe53f8d" (UID: "7a1d264d-7e6b-4910-9656-61ffefe53f8d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.476357 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a1d264d-7e6b-4910-9656-61ffefe53f8d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a1d264d-7e6b-4910-9656-61ffefe53f8d" (UID: "7a1d264d-7e6b-4910-9656-61ffefe53f8d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.487982 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.539177 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a1d264d-7e6b-4910-9656-61ffefe53f8d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.539205 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a1d264d-7e6b-4910-9656-61ffefe53f8d-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:41 crc kubenswrapper[4696]: I0321 08:49:41.539215 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4s5z\" (UniqueName: \"kubernetes.io/projected/7a1d264d-7e6b-4910-9656-61ffefe53f8d-kube-api-access-r4s5z\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:41.964360 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.074792 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267","Type":"ContainerStarted","Data":"b8241fdbacc9c7a2b134cea9bc3b5ff15242ba911ba9a0af4a06edcfb5ba33ef"} Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.075623 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.084113 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7a1d264d-7e6b-4910-9656-61ffefe53f8d","Type":"ContainerDied","Data":"6fc482dfefa52fc80cb91ce841922647c1ee5e44e9d07ec623c7fadf7e6e19a4"} Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.084157 4696 scope.go:117] "RemoveContainer" containerID="92ad6b8cfc9519d7f9031eec9efa261d2fe191665ac15c9fca809d3ecc7e6279" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.084200 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.104260 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.104246871 podStartE2EDuration="2.104246871s" podCreationTimestamp="2026-03-21 08:49:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:49:42.103342835 +0000 UTC m=+1316.224223548" watchObservedRunningTime="2026-03-21 08:49:42.104246871 +0000 UTC m=+1316.225127584" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.181289 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.201620 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.221805 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 08:49:42 crc kubenswrapper[4696]: E0321 08:49:42.222343 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a1d264d-7e6b-4910-9656-61ffefe53f8d" containerName="nova-scheduler-scheduler" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.222363 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a1d264d-7e6b-4910-9656-61ffefe53f8d" containerName="nova-scheduler-scheduler" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.222585 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a1d264d-7e6b-4910-9656-61ffefe53f8d" containerName="nova-scheduler-scheduler" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.223441 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.226128 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.238495 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.381370 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/910aaf27-d4d5-4031-9971-95b472f065d8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"910aaf27-d4d5-4031-9971-95b472f065d8\") " pod="openstack/nova-scheduler-0" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.381690 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/910aaf27-d4d5-4031-9971-95b472f065d8-config-data\") pod \"nova-scheduler-0\" (UID: \"910aaf27-d4d5-4031-9971-95b472f065d8\") " pod="openstack/nova-scheduler-0" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.381729 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4748t\" (UniqueName: \"kubernetes.io/projected/910aaf27-d4d5-4031-9971-95b472f065d8-kube-api-access-4748t\") pod \"nova-scheduler-0\" (UID: \"910aaf27-d4d5-4031-9971-95b472f065d8\") " pod="openstack/nova-scheduler-0" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.483672 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/910aaf27-d4d5-4031-9971-95b472f065d8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"910aaf27-d4d5-4031-9971-95b472f065d8\") " pod="openstack/nova-scheduler-0" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.484239 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/910aaf27-d4d5-4031-9971-95b472f065d8-config-data\") pod \"nova-scheduler-0\" (UID: \"910aaf27-d4d5-4031-9971-95b472f065d8\") " pod="openstack/nova-scheduler-0" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.484490 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4748t\" (UniqueName: \"kubernetes.io/projected/910aaf27-d4d5-4031-9971-95b472f065d8-kube-api-access-4748t\") pod \"nova-scheduler-0\" (UID: \"910aaf27-d4d5-4031-9971-95b472f065d8\") " pod="openstack/nova-scheduler-0" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.487671 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/910aaf27-d4d5-4031-9971-95b472f065d8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"910aaf27-d4d5-4031-9971-95b472f065d8\") " pod="openstack/nova-scheduler-0" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.487802 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/910aaf27-d4d5-4031-9971-95b472f065d8-config-data\") pod \"nova-scheduler-0\" (UID: \"910aaf27-d4d5-4031-9971-95b472f065d8\") " pod="openstack/nova-scheduler-0" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.502370 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4748t\" (UniqueName: \"kubernetes.io/projected/910aaf27-d4d5-4031-9971-95b472f065d8-kube-api-access-4748t\") pod \"nova-scheduler-0\" (UID: \"910aaf27-d4d5-4031-9971-95b472f065d8\") " pod="openstack/nova-scheduler-0" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.551802 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a1d264d-7e6b-4910-9656-61ffefe53f8d" path="/var/lib/kubelet/pods/7a1d264d-7e6b-4910-9656-61ffefe53f8d/volumes" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.553243 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab19b72e-ead0-4baa-ba2a-badfae78ffa0" path="/var/lib/kubelet/pods/ab19b72e-ead0-4baa-ba2a-badfae78ffa0/volumes" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.568800 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.689511 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 21 08:49:42 crc kubenswrapper[4696]: I0321 08:49:42.689542 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 21 08:49:43 crc kubenswrapper[4696]: I0321 08:49:43.102065 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb5d65be-1467-4ef3-9a33-c7707c4b23d2","Type":"ContainerStarted","Data":"14fa5fb139c9f816fc8709e2ea4371101baea2bfb37b5e603c3c80787b9c3fbd"} Mar 21 08:49:43 crc kubenswrapper[4696]: I0321 08:49:43.102425 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb5d65be-1467-4ef3-9a33-c7707c4b23d2","Type":"ContainerStarted","Data":"bcdfb356288d9965ba7a3b0a90fc27796a87029add34df6a18564f8e4ff98898"} Mar 21 08:49:43 crc kubenswrapper[4696]: I0321 08:49:43.102455 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb5d65be-1467-4ef3-9a33-c7707c4b23d2","Type":"ContainerStarted","Data":"bba16a65a33465580d3dfe23a3df54d414630960638aa0c79ced29884712d231"} Mar 21 08:49:43 crc kubenswrapper[4696]: I0321 08:49:43.108390 4696 generic.go:334] "Generic (PLEG): container finished" podID="34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1" containerID="73da593316c9714471379b39652eebf6cf072a108dd7bbee42f1ef069967be5f" exitCode=0 Mar 21 08:49:43 crc kubenswrapper[4696]: I0321 08:49:43.108509 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1","Type":"ContainerDied","Data":"73da593316c9714471379b39652eebf6cf072a108dd7bbee42f1ef069967be5f"} Mar 21 08:49:43 crc kubenswrapper[4696]: I0321 08:49:43.143731 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.143705926 podStartE2EDuration="2.143705926s" podCreationTimestamp="2026-03-21 08:49:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:49:43.123936978 +0000 UTC m=+1317.244817691" watchObservedRunningTime="2026-03-21 08:49:43.143705926 +0000 UTC m=+1317.264586639" Mar 21 08:49:43 crc kubenswrapper[4696]: I0321 08:49:43.435881 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 08:49:43 crc kubenswrapper[4696]: I0321 08:49:43.764460 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 08:49:43 crc kubenswrapper[4696]: I0321 08:49:43.935709 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9lpt\" (UniqueName: \"kubernetes.io/projected/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-kube-api-access-s9lpt\") pod \"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1\" (UID: \"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1\") " Mar 21 08:49:43 crc kubenswrapper[4696]: I0321 08:49:43.935796 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-logs\") pod \"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1\" (UID: \"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1\") " Mar 21 08:49:43 crc kubenswrapper[4696]: I0321 08:49:43.936050 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-config-data\") pod \"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1\" (UID: \"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1\") " Mar 21 08:49:43 crc kubenswrapper[4696]: I0321 08:49:43.936099 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-combined-ca-bundle\") pod \"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1\" (UID: \"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1\") " Mar 21 08:49:43 crc kubenswrapper[4696]: I0321 08:49:43.936341 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-logs" (OuterVolumeSpecName: "logs") pod "34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1" (UID: "34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:49:43 crc kubenswrapper[4696]: I0321 08:49:43.936833 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-logs\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:43 crc kubenswrapper[4696]: I0321 08:49:43.943558 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-kube-api-access-s9lpt" (OuterVolumeSpecName: "kube-api-access-s9lpt") pod "34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1" (UID: "34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1"). InnerVolumeSpecName "kube-api-access-s9lpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:49:43 crc kubenswrapper[4696]: I0321 08:49:43.965951 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1" (UID: "34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:43 crc kubenswrapper[4696]: I0321 08:49:43.984108 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-config-data" (OuterVolumeSpecName: "config-data") pod "34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1" (UID: "34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.038195 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.038227 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.038238 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9lpt\" (UniqueName: \"kubernetes.io/projected/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1-kube-api-access-s9lpt\") on node \"crc\" DevicePath \"\"" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.121902 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1","Type":"ContainerDied","Data":"7cbc7eeb169246f38abd7ccd436eee26e9fd9cf70f9f5628d8f6031403cf0f9c"} Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.121952 4696 scope.go:117] "RemoveContainer" containerID="73da593316c9714471379b39652eebf6cf072a108dd7bbee42f1ef069967be5f" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.121921 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.123945 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"910aaf27-d4d5-4031-9971-95b472f065d8","Type":"ContainerStarted","Data":"1cae17529197bff02832d32292e43ac560bf06d48c328a2323d64f5a3646602d"} Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.123990 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"910aaf27-d4d5-4031-9971-95b472f065d8","Type":"ContainerStarted","Data":"19e6f747772077f00d1425247e124bf2530bdaf1860b5cee3cbd1a69eeed7529"} Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.148799 4696 scope.go:117] "RemoveContainer" containerID="a5078c0880aedda106502b88c3ed962c04f889b9117d6af7d22c03cccf1fc203" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.152742 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.152725677 podStartE2EDuration="2.152725677s" podCreationTimestamp="2026-03-21 08:49:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:49:44.144510509 +0000 UTC m=+1318.265391222" watchObservedRunningTime="2026-03-21 08:49:44.152725677 +0000 UTC m=+1318.273606390" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.177808 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.196831 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.213086 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 21 08:49:44 crc kubenswrapper[4696]: E0321 08:49:44.213509 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1" containerName="nova-api-api" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.213525 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1" containerName="nova-api-api" Mar 21 08:49:44 crc kubenswrapper[4696]: E0321 08:49:44.213535 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1" containerName="nova-api-log" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.213541 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1" containerName="nova-api-log" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.213738 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1" containerName="nova-api-log" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.213764 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1" containerName="nova-api-api" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.214906 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.216788 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.225514 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.242336 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tplb\" (UniqueName: \"kubernetes.io/projected/2bbaeade-9a94-419e-8a55-3b90334ea861-kube-api-access-8tplb\") pod \"nova-api-0\" (UID: \"2bbaeade-9a94-419e-8a55-3b90334ea861\") " pod="openstack/nova-api-0" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.242403 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bbaeade-9a94-419e-8a55-3b90334ea861-config-data\") pod \"nova-api-0\" (UID: \"2bbaeade-9a94-419e-8a55-3b90334ea861\") " pod="openstack/nova-api-0" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.242430 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bbaeade-9a94-419e-8a55-3b90334ea861-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2bbaeade-9a94-419e-8a55-3b90334ea861\") " pod="openstack/nova-api-0" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.242495 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bbaeade-9a94-419e-8a55-3b90334ea861-logs\") pod \"nova-api-0\" (UID: \"2bbaeade-9a94-419e-8a55-3b90334ea861\") " pod="openstack/nova-api-0" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.345527 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tplb\" (UniqueName: \"kubernetes.io/projected/2bbaeade-9a94-419e-8a55-3b90334ea861-kube-api-access-8tplb\") pod \"nova-api-0\" (UID: \"2bbaeade-9a94-419e-8a55-3b90334ea861\") " pod="openstack/nova-api-0" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.345638 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bbaeade-9a94-419e-8a55-3b90334ea861-config-data\") pod \"nova-api-0\" (UID: \"2bbaeade-9a94-419e-8a55-3b90334ea861\") " pod="openstack/nova-api-0" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.345674 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bbaeade-9a94-419e-8a55-3b90334ea861-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2bbaeade-9a94-419e-8a55-3b90334ea861\") " pod="openstack/nova-api-0" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.345739 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bbaeade-9a94-419e-8a55-3b90334ea861-logs\") pod \"nova-api-0\" (UID: \"2bbaeade-9a94-419e-8a55-3b90334ea861\") " pod="openstack/nova-api-0" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.346283 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bbaeade-9a94-419e-8a55-3b90334ea861-logs\") pod \"nova-api-0\" (UID: \"2bbaeade-9a94-419e-8a55-3b90334ea861\") " pod="openstack/nova-api-0" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.349551 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bbaeade-9a94-419e-8a55-3b90334ea861-config-data\") pod \"nova-api-0\" (UID: \"2bbaeade-9a94-419e-8a55-3b90334ea861\") " pod="openstack/nova-api-0" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.350067 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bbaeade-9a94-419e-8a55-3b90334ea861-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2bbaeade-9a94-419e-8a55-3b90334ea861\") " pod="openstack/nova-api-0" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.362181 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tplb\" (UniqueName: \"kubernetes.io/projected/2bbaeade-9a94-419e-8a55-3b90334ea861-kube-api-access-8tplb\") pod \"nova-api-0\" (UID: \"2bbaeade-9a94-419e-8a55-3b90334ea861\") " pod="openstack/nova-api-0" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.531214 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 08:49:44 crc kubenswrapper[4696]: I0321 08:49:44.549986 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1" path="/var/lib/kubelet/pods/34c41bd7-ac72-43e0-a96a-3b7ec06e0fd1/volumes" Mar 21 08:49:45 crc kubenswrapper[4696]: I0321 08:49:45.006985 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 21 08:49:45 crc kubenswrapper[4696]: I0321 08:49:45.135165 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2bbaeade-9a94-419e-8a55-3b90334ea861","Type":"ContainerStarted","Data":"a78d61dbd40752f1465b0a38d1c32371c57aa9fa2177c45af9226ee4e300c5d9"} Mar 21 08:49:46 crc kubenswrapper[4696]: I0321 08:49:46.149887 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2bbaeade-9a94-419e-8a55-3b90334ea861","Type":"ContainerStarted","Data":"b38d446de9a50a7f1fa4339e2e05e5dcafd2d9b76fa6f8217ea05a80efc71616"} Mar 21 08:49:46 crc kubenswrapper[4696]: I0321 08:49:46.150213 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2bbaeade-9a94-419e-8a55-3b90334ea861","Type":"ContainerStarted","Data":"92865c95ba3fbb05da9e233bda7f92f1886c5f60e43902ac526e81c361e58bd6"} Mar 21 08:49:46 crc kubenswrapper[4696]: I0321 08:49:46.181543 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.181522118 podStartE2EDuration="2.181522118s" podCreationTimestamp="2026-03-21 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:49:46.16894914 +0000 UTC m=+1320.289829853" watchObservedRunningTime="2026-03-21 08:49:46.181522118 +0000 UTC m=+1320.302402821" Mar 21 08:49:47 crc kubenswrapper[4696]: I0321 08:49:47.569472 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 21 08:49:50 crc kubenswrapper[4696]: I0321 08:49:50.465325 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Mar 21 08:49:51 crc kubenswrapper[4696]: I0321 08:49:51.488767 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 21 08:49:51 crc kubenswrapper[4696]: I0321 08:49:51.489155 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 21 08:49:52 crc kubenswrapper[4696]: I0321 08:49:52.501018 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fb5d65be-1467-4ef3-9a33-c7707c4b23d2" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.227:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 21 08:49:52 crc kubenswrapper[4696]: I0321 08:49:52.501032 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fb5d65be-1467-4ef3-9a33-c7707c4b23d2" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.227:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 21 08:49:52 crc kubenswrapper[4696]: I0321 08:49:52.569210 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 21 08:49:52 crc kubenswrapper[4696]: I0321 08:49:52.596552 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 21 08:49:53 crc kubenswrapper[4696]: I0321 08:49:53.245362 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 21 08:49:54 crc kubenswrapper[4696]: I0321 08:49:54.531490 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 21 08:49:54 crc kubenswrapper[4696]: I0321 08:49:54.531550 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 21 08:49:55 crc kubenswrapper[4696]: I0321 08:49:55.614086 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2bbaeade-9a94-419e-8a55-3b90334ea861" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.229:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 21 08:49:55 crc kubenswrapper[4696]: I0321 08:49:55.614121 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2bbaeade-9a94-419e-8a55-3b90334ea861" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.229:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 21 08:49:59 crc kubenswrapper[4696]: I0321 08:49:59.488535 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 21 08:49:59 crc kubenswrapper[4696]: I0321 08:49:59.488795 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 21 08:50:00 crc kubenswrapper[4696]: I0321 08:50:00.147130 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568050-g4f92"] Mar 21 08:50:00 crc kubenswrapper[4696]: I0321 08:50:00.148895 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568050-g4f92" Mar 21 08:50:00 crc kubenswrapper[4696]: I0321 08:50:00.151661 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 08:50:00 crc kubenswrapper[4696]: I0321 08:50:00.151943 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 08:50:00 crc kubenswrapper[4696]: I0321 08:50:00.152113 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 08:50:00 crc kubenswrapper[4696]: I0321 08:50:00.159838 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568050-g4f92"] Mar 21 08:50:00 crc kubenswrapper[4696]: I0321 08:50:00.191161 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qsjw\" (UniqueName: \"kubernetes.io/projected/7e8563c8-4723-4e92-9972-3cddb4c529e7-kube-api-access-9qsjw\") pod \"auto-csr-approver-29568050-g4f92\" (UID: \"7e8563c8-4723-4e92-9972-3cddb4c529e7\") " pod="openshift-infra/auto-csr-approver-29568050-g4f92" Mar 21 08:50:00 crc kubenswrapper[4696]: I0321 08:50:00.293684 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qsjw\" (UniqueName: \"kubernetes.io/projected/7e8563c8-4723-4e92-9972-3cddb4c529e7-kube-api-access-9qsjw\") pod \"auto-csr-approver-29568050-g4f92\" (UID: \"7e8563c8-4723-4e92-9972-3cddb4c529e7\") " pod="openshift-infra/auto-csr-approver-29568050-g4f92" Mar 21 08:50:00 crc kubenswrapper[4696]: I0321 08:50:00.315150 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qsjw\" (UniqueName: \"kubernetes.io/projected/7e8563c8-4723-4e92-9972-3cddb4c529e7-kube-api-access-9qsjw\") pod \"auto-csr-approver-29568050-g4f92\" (UID: \"7e8563c8-4723-4e92-9972-3cddb4c529e7\") " pod="openshift-infra/auto-csr-approver-29568050-g4f92" Mar 21 08:50:00 crc kubenswrapper[4696]: I0321 08:50:00.466927 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568050-g4f92" Mar 21 08:50:01 crc kubenswrapper[4696]: I0321 08:50:01.043367 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568050-g4f92"] Mar 21 08:50:01 crc kubenswrapper[4696]: I0321 08:50:01.285484 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568050-g4f92" event={"ID":"7e8563c8-4723-4e92-9972-3cddb4c529e7","Type":"ContainerStarted","Data":"090ab38a47b070516d6bd3afe4927cf8545a71a33f748ff4b2846d8a79a367c1"} Mar 21 08:50:01 crc kubenswrapper[4696]: I0321 08:50:01.493506 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 21 08:50:01 crc kubenswrapper[4696]: I0321 08:50:01.494754 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 21 08:50:01 crc kubenswrapper[4696]: I0321 08:50:01.498264 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 21 08:50:02 crc kubenswrapper[4696]: I0321 08:50:02.301751 4696 generic.go:334] "Generic (PLEG): container finished" podID="0b436237-c742-46e2-a4b0-f0eec79d9fd2" containerID="68a26d8e09f41c2c1a6c42fa9af0f87c665effab807593d28764e1852ec9d521" exitCode=137 Mar 21 08:50:02 crc kubenswrapper[4696]: I0321 08:50:02.301861 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0b436237-c742-46e2-a4b0-f0eec79d9fd2","Type":"ContainerDied","Data":"68a26d8e09f41c2c1a6c42fa9af0f87c665effab807593d28764e1852ec9d521"} Mar 21 08:50:02 crc kubenswrapper[4696]: I0321 08:50:02.313314 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 21 08:50:02 crc kubenswrapper[4696]: I0321 08:50:02.526285 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:02 crc kubenswrapper[4696]: I0321 08:50:02.531956 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 21 08:50:02 crc kubenswrapper[4696]: I0321 08:50:02.531999 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 21 08:50:02 crc kubenswrapper[4696]: I0321 08:50:02.553110 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwwh9\" (UniqueName: \"kubernetes.io/projected/0b436237-c742-46e2-a4b0-f0eec79d9fd2-kube-api-access-bwwh9\") pod \"0b436237-c742-46e2-a4b0-f0eec79d9fd2\" (UID: \"0b436237-c742-46e2-a4b0-f0eec79d9fd2\") " Mar 21 08:50:02 crc kubenswrapper[4696]: I0321 08:50:02.553218 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b436237-c742-46e2-a4b0-f0eec79d9fd2-combined-ca-bundle\") pod \"0b436237-c742-46e2-a4b0-f0eec79d9fd2\" (UID: \"0b436237-c742-46e2-a4b0-f0eec79d9fd2\") " Mar 21 08:50:02 crc kubenswrapper[4696]: I0321 08:50:02.553327 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b436237-c742-46e2-a4b0-f0eec79d9fd2-config-data\") pod \"0b436237-c742-46e2-a4b0-f0eec79d9fd2\" (UID: \"0b436237-c742-46e2-a4b0-f0eec79d9fd2\") " Mar 21 08:50:02 crc kubenswrapper[4696]: I0321 08:50:02.569068 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b436237-c742-46e2-a4b0-f0eec79d9fd2-kube-api-access-bwwh9" (OuterVolumeSpecName: "kube-api-access-bwwh9") pod "0b436237-c742-46e2-a4b0-f0eec79d9fd2" (UID: "0b436237-c742-46e2-a4b0-f0eec79d9fd2"). InnerVolumeSpecName "kube-api-access-bwwh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:50:02 crc kubenswrapper[4696]: I0321 08:50:02.591383 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b436237-c742-46e2-a4b0-f0eec79d9fd2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b436237-c742-46e2-a4b0-f0eec79d9fd2" (UID: "0b436237-c742-46e2-a4b0-f0eec79d9fd2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:02 crc kubenswrapper[4696]: I0321 08:50:02.623023 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b436237-c742-46e2-a4b0-f0eec79d9fd2-config-data" (OuterVolumeSpecName: "config-data") pod "0b436237-c742-46e2-a4b0-f0eec79d9fd2" (UID: "0b436237-c742-46e2-a4b0-f0eec79d9fd2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:02 crc kubenswrapper[4696]: I0321 08:50:02.656213 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwwh9\" (UniqueName: \"kubernetes.io/projected/0b436237-c742-46e2-a4b0-f0eec79d9fd2-kube-api-access-bwwh9\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:02 crc kubenswrapper[4696]: I0321 08:50:02.656254 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b436237-c742-46e2-a4b0-f0eec79d9fd2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:02 crc kubenswrapper[4696]: I0321 08:50:02.656266 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b436237-c742-46e2-a4b0-f0eec79d9fd2-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.154238 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.330793 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568050-g4f92" event={"ID":"7e8563c8-4723-4e92-9972-3cddb4c529e7","Type":"ContainerStarted","Data":"0731f21de16f79a7267e66eb9659e169aec45ac89300ab26801aeb5d1d5d7aea"} Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.334701 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0b436237-c742-46e2-a4b0-f0eec79d9fd2","Type":"ContainerDied","Data":"4aa2117782eac7109f7fb6352e659bd7322965a6a6b3f0e92fb4a5df746bccf9"} Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.334711 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.334762 4696 scope.go:117] "RemoveContainer" containerID="68a26d8e09f41c2c1a6c42fa9af0f87c665effab807593d28764e1852ec9d521" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.352508 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29568050-g4f92" podStartSLOduration=2.206898399 podStartE2EDuration="3.352483953s" podCreationTimestamp="2026-03-21 08:50:00 +0000 UTC" firstStartedPulling="2026-03-21 08:50:01.065300506 +0000 UTC m=+1335.186181209" lastFinishedPulling="2026-03-21 08:50:02.21088605 +0000 UTC m=+1336.331766763" observedRunningTime="2026-03-21 08:50:03.343782092 +0000 UTC m=+1337.464662805" watchObservedRunningTime="2026-03-21 08:50:03.352483953 +0000 UTC m=+1337.473364666" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.378187 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.388139 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.400122 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 21 08:50:03 crc kubenswrapper[4696]: E0321 08:50:03.400560 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b436237-c742-46e2-a4b0-f0eec79d9fd2" containerName="nova-cell1-novncproxy-novncproxy" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.400580 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b436237-c742-46e2-a4b0-f0eec79d9fd2" containerName="nova-cell1-novncproxy-novncproxy" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.400801 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b436237-c742-46e2-a4b0-f0eec79d9fd2" containerName="nova-cell1-novncproxy-novncproxy" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.401468 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.403795 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.403893 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.405949 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.425562 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.474104 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f424006b-438e-46cc-8d5f-70d8d0346f40-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f424006b-438e-46cc-8d5f-70d8d0346f40\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.474213 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f424006b-438e-46cc-8d5f-70d8d0346f40-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f424006b-438e-46cc-8d5f-70d8d0346f40\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.474290 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f424006b-438e-46cc-8d5f-70d8d0346f40-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f424006b-438e-46cc-8d5f-70d8d0346f40\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.474349 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lrtz\" (UniqueName: \"kubernetes.io/projected/f424006b-438e-46cc-8d5f-70d8d0346f40-kube-api-access-6lrtz\") pod \"nova-cell1-novncproxy-0\" (UID: \"f424006b-438e-46cc-8d5f-70d8d0346f40\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.474408 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f424006b-438e-46cc-8d5f-70d8d0346f40-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f424006b-438e-46cc-8d5f-70d8d0346f40\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.576019 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lrtz\" (UniqueName: \"kubernetes.io/projected/f424006b-438e-46cc-8d5f-70d8d0346f40-kube-api-access-6lrtz\") pod \"nova-cell1-novncproxy-0\" (UID: \"f424006b-438e-46cc-8d5f-70d8d0346f40\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.576131 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f424006b-438e-46cc-8d5f-70d8d0346f40-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f424006b-438e-46cc-8d5f-70d8d0346f40\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.576179 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f424006b-438e-46cc-8d5f-70d8d0346f40-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f424006b-438e-46cc-8d5f-70d8d0346f40\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.576238 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f424006b-438e-46cc-8d5f-70d8d0346f40-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f424006b-438e-46cc-8d5f-70d8d0346f40\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.576329 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f424006b-438e-46cc-8d5f-70d8d0346f40-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f424006b-438e-46cc-8d5f-70d8d0346f40\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.585346 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f424006b-438e-46cc-8d5f-70d8d0346f40-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f424006b-438e-46cc-8d5f-70d8d0346f40\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.585344 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f424006b-438e-46cc-8d5f-70d8d0346f40-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f424006b-438e-46cc-8d5f-70d8d0346f40\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.585518 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f424006b-438e-46cc-8d5f-70d8d0346f40-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f424006b-438e-46cc-8d5f-70d8d0346f40\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.585616 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f424006b-438e-46cc-8d5f-70d8d0346f40-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f424006b-438e-46cc-8d5f-70d8d0346f40\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.605134 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lrtz\" (UniqueName: \"kubernetes.io/projected/f424006b-438e-46cc-8d5f-70d8d0346f40-kube-api-access-6lrtz\") pod \"nova-cell1-novncproxy-0\" (UID: \"f424006b-438e-46cc-8d5f-70d8d0346f40\") " pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:03 crc kubenswrapper[4696]: I0321 08:50:03.718133 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:04 crc kubenswrapper[4696]: I0321 08:50:04.216270 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 21 08:50:04 crc kubenswrapper[4696]: I0321 08:50:04.347107 4696 generic.go:334] "Generic (PLEG): container finished" podID="7e8563c8-4723-4e92-9972-3cddb4c529e7" containerID="0731f21de16f79a7267e66eb9659e169aec45ac89300ab26801aeb5d1d5d7aea" exitCode=0 Mar 21 08:50:04 crc kubenswrapper[4696]: I0321 08:50:04.347159 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568050-g4f92" event={"ID":"7e8563c8-4723-4e92-9972-3cddb4c529e7","Type":"ContainerDied","Data":"0731f21de16f79a7267e66eb9659e169aec45ac89300ab26801aeb5d1d5d7aea"} Mar 21 08:50:04 crc kubenswrapper[4696]: I0321 08:50:04.348284 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f424006b-438e-46cc-8d5f-70d8d0346f40","Type":"ContainerStarted","Data":"4a70da6b73416316df5801b09f7d8d17fe7691024bfb77e5ecc9f878b34ca8fd"} Mar 21 08:50:04 crc kubenswrapper[4696]: I0321 08:50:04.545714 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b436237-c742-46e2-a4b0-f0eec79d9fd2" path="/var/lib/kubelet/pods/0b436237-c742-46e2-a4b0-f0eec79d9fd2/volumes" Mar 21 08:50:04 crc kubenswrapper[4696]: I0321 08:50:04.546327 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 21 08:50:04 crc kubenswrapper[4696]: I0321 08:50:04.550652 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 21 08:50:04 crc kubenswrapper[4696]: I0321 08:50:04.552398 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.359805 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f424006b-438e-46cc-8d5f-70d8d0346f40","Type":"ContainerStarted","Data":"8027fde5ab885c8bffbeafdc648e527b4b1884cd287e53de2491f43eb92787b6"} Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.365567 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.383782 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.383758892 podStartE2EDuration="2.383758892s" podCreationTimestamp="2026-03-21 08:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:50:05.382037134 +0000 UTC m=+1339.502917847" watchObservedRunningTime="2026-03-21 08:50:05.383758892 +0000 UTC m=+1339.504639605" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.592029 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5fd9b586ff-xr8tz"] Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.594254 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.607065 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5fd9b586ff-xr8tz"] Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.720397 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-dns-swift-storage-0\") pod \"dnsmasq-dns-5fd9b586ff-xr8tz\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.720461 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-config\") pod \"dnsmasq-dns-5fd9b586ff-xr8tz\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.720663 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-dns-svc\") pod \"dnsmasq-dns-5fd9b586ff-xr8tz\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.720719 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-ovsdbserver-sb\") pod \"dnsmasq-dns-5fd9b586ff-xr8tz\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.720766 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sllqw\" (UniqueName: \"kubernetes.io/projected/4acec0f5-fbf7-4900-92aa-181b72767e79-kube-api-access-sllqw\") pod \"dnsmasq-dns-5fd9b586ff-xr8tz\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.721196 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-ovsdbserver-nb\") pod \"dnsmasq-dns-5fd9b586ff-xr8tz\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.824192 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-ovsdbserver-nb\") pod \"dnsmasq-dns-5fd9b586ff-xr8tz\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.824274 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-dns-swift-storage-0\") pod \"dnsmasq-dns-5fd9b586ff-xr8tz\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.824306 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-config\") pod \"dnsmasq-dns-5fd9b586ff-xr8tz\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.824355 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-dns-svc\") pod \"dnsmasq-dns-5fd9b586ff-xr8tz\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.824371 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-ovsdbserver-sb\") pod \"dnsmasq-dns-5fd9b586ff-xr8tz\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.824388 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sllqw\" (UniqueName: \"kubernetes.io/projected/4acec0f5-fbf7-4900-92aa-181b72767e79-kube-api-access-sllqw\") pod \"dnsmasq-dns-5fd9b586ff-xr8tz\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.825488 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-ovsdbserver-nb\") pod \"dnsmasq-dns-5fd9b586ff-xr8tz\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.825512 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-config\") pod \"dnsmasq-dns-5fd9b586ff-xr8tz\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.825695 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-dns-svc\") pod \"dnsmasq-dns-5fd9b586ff-xr8tz\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.827202 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-ovsdbserver-sb\") pod \"dnsmasq-dns-5fd9b586ff-xr8tz\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.839400 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-dns-swift-storage-0\") pod \"dnsmasq-dns-5fd9b586ff-xr8tz\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.846533 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sllqw\" (UniqueName: \"kubernetes.io/projected/4acec0f5-fbf7-4900-92aa-181b72767e79-kube-api-access-sllqw\") pod \"dnsmasq-dns-5fd9b586ff-xr8tz\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:05 crc kubenswrapper[4696]: I0321 08:50:05.920860 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:06 crc kubenswrapper[4696]: I0321 08:50:06.179435 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568050-g4f92" Mar 21 08:50:06 crc kubenswrapper[4696]: I0321 08:50:06.335281 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qsjw\" (UniqueName: \"kubernetes.io/projected/7e8563c8-4723-4e92-9972-3cddb4c529e7-kube-api-access-9qsjw\") pod \"7e8563c8-4723-4e92-9972-3cddb4c529e7\" (UID: \"7e8563c8-4723-4e92-9972-3cddb4c529e7\") " Mar 21 08:50:06 crc kubenswrapper[4696]: I0321 08:50:06.343234 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e8563c8-4723-4e92-9972-3cddb4c529e7-kube-api-access-9qsjw" (OuterVolumeSpecName: "kube-api-access-9qsjw") pod "7e8563c8-4723-4e92-9972-3cddb4c529e7" (UID: "7e8563c8-4723-4e92-9972-3cddb4c529e7"). InnerVolumeSpecName "kube-api-access-9qsjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:50:06 crc kubenswrapper[4696]: I0321 08:50:06.399411 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568050-g4f92" event={"ID":"7e8563c8-4723-4e92-9972-3cddb4c529e7","Type":"ContainerDied","Data":"090ab38a47b070516d6bd3afe4927cf8545a71a33f748ff4b2846d8a79a367c1"} Mar 21 08:50:06 crc kubenswrapper[4696]: I0321 08:50:06.399448 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="090ab38a47b070516d6bd3afe4927cf8545a71a33f748ff4b2846d8a79a367c1" Mar 21 08:50:06 crc kubenswrapper[4696]: I0321 08:50:06.399523 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568050-g4f92" Mar 21 08:50:06 crc kubenswrapper[4696]: I0321 08:50:06.431859 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568044-g82fg"] Mar 21 08:50:06 crc kubenswrapper[4696]: I0321 08:50:06.440113 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qsjw\" (UniqueName: \"kubernetes.io/projected/7e8563c8-4723-4e92-9972-3cddb4c529e7-kube-api-access-9qsjw\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:06 crc kubenswrapper[4696]: I0321 08:50:06.445222 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568044-g82fg"] Mar 21 08:50:06 crc kubenswrapper[4696]: I0321 08:50:06.506113 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5fd9b586ff-xr8tz"] Mar 21 08:50:06 crc kubenswrapper[4696]: I0321 08:50:06.553872 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="703e51f2-bf70-45c9-ac92-3d96a4d49d18" path="/var/lib/kubelet/pods/703e51f2-bf70-45c9-ac92-3d96a4d49d18/volumes" Mar 21 08:50:07 crc kubenswrapper[4696]: I0321 08:50:07.457600 4696 generic.go:334] "Generic (PLEG): container finished" podID="4acec0f5-fbf7-4900-92aa-181b72767e79" containerID="4c7c5acb31fa5b1d43fabd39637850ce34e084b6c8ea6c6ebf1370f8dacaef67" exitCode=0 Mar 21 08:50:07 crc kubenswrapper[4696]: I0321 08:50:07.458583 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" event={"ID":"4acec0f5-fbf7-4900-92aa-181b72767e79","Type":"ContainerDied","Data":"4c7c5acb31fa5b1d43fabd39637850ce34e084b6c8ea6c6ebf1370f8dacaef67"} Mar 21 08:50:07 crc kubenswrapper[4696]: I0321 08:50:07.458690 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" event={"ID":"4acec0f5-fbf7-4900-92aa-181b72767e79","Type":"ContainerStarted","Data":"69f6eba6b1c8ebe80e384b5f2ef91f9adfb73a525452be2737c44329e6fbc632"} Mar 21 08:50:08 crc kubenswrapper[4696]: I0321 08:50:08.319324 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 21 08:50:08 crc kubenswrapper[4696]: I0321 08:50:08.479180 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" event={"ID":"4acec0f5-fbf7-4900-92aa-181b72767e79","Type":"ContainerStarted","Data":"98c8fe9d6eb7efe4cbb38d9718d2575729319445cfe0682e2f960ac47a953850"} Mar 21 08:50:08 crc kubenswrapper[4696]: I0321 08:50:08.479617 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:08 crc kubenswrapper[4696]: I0321 08:50:08.480696 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2bbaeade-9a94-419e-8a55-3b90334ea861" containerName="nova-api-api" containerID="cri-o://b38d446de9a50a7f1fa4339e2e05e5dcafd2d9b76fa6f8217ea05a80efc71616" gracePeriod=30 Mar 21 08:50:08 crc kubenswrapper[4696]: I0321 08:50:08.480635 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2bbaeade-9a94-419e-8a55-3b90334ea861" containerName="nova-api-log" containerID="cri-o://92865c95ba3fbb05da9e233bda7f92f1886c5f60e43902ac526e81c361e58bd6" gracePeriod=30 Mar 21 08:50:08 crc kubenswrapper[4696]: I0321 08:50:08.511281 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" podStartSLOduration=3.511260118 podStartE2EDuration="3.511260118s" podCreationTimestamp="2026-03-21 08:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:50:08.499840672 +0000 UTC m=+1342.620721385" watchObservedRunningTime="2026-03-21 08:50:08.511260118 +0000 UTC m=+1342.632140841" Mar 21 08:50:08 crc kubenswrapper[4696]: I0321 08:50:08.719245 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:09 crc kubenswrapper[4696]: I0321 08:50:09.288132 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:50:09 crc kubenswrapper[4696]: I0321 08:50:09.288481 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e1df5e20-2002-4b15-a6b6-f23874343914" containerName="ceilometer-central-agent" containerID="cri-o://bece433f574d7f51b9470e8875dfab386d4755bbb68df8962d52b36c582ba420" gracePeriod=30 Mar 21 08:50:09 crc kubenswrapper[4696]: I0321 08:50:09.288531 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e1df5e20-2002-4b15-a6b6-f23874343914" containerName="proxy-httpd" containerID="cri-o://52712668b4c83dc1d0113247da7073898d3f4a1431b21f117f82c1a4d0118d02" gracePeriod=30 Mar 21 08:50:09 crc kubenswrapper[4696]: I0321 08:50:09.288595 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e1df5e20-2002-4b15-a6b6-f23874343914" containerName="ceilometer-notification-agent" containerID="cri-o://7eb1093e5d953b48d0bcdd3d193e4c4a05d3dd7fc4246704a7cee0195e1cb843" gracePeriod=30 Mar 21 08:50:09 crc kubenswrapper[4696]: I0321 08:50:09.288602 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e1df5e20-2002-4b15-a6b6-f23874343914" containerName="sg-core" containerID="cri-o://7fbc416c207de813688566084124df0e65f347c6de903ae1dcdf3d7a33fdb3e4" gracePeriod=30 Mar 21 08:50:09 crc kubenswrapper[4696]: I0321 08:50:09.491244 4696 generic.go:334] "Generic (PLEG): container finished" podID="e1df5e20-2002-4b15-a6b6-f23874343914" containerID="7fbc416c207de813688566084124df0e65f347c6de903ae1dcdf3d7a33fdb3e4" exitCode=2 Mar 21 08:50:09 crc kubenswrapper[4696]: I0321 08:50:09.491295 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1df5e20-2002-4b15-a6b6-f23874343914","Type":"ContainerDied","Data":"7fbc416c207de813688566084124df0e65f347c6de903ae1dcdf3d7a33fdb3e4"} Mar 21 08:50:09 crc kubenswrapper[4696]: I0321 08:50:09.494105 4696 generic.go:334] "Generic (PLEG): container finished" podID="2bbaeade-9a94-419e-8a55-3b90334ea861" containerID="92865c95ba3fbb05da9e233bda7f92f1886c5f60e43902ac526e81c361e58bd6" exitCode=143 Mar 21 08:50:09 crc kubenswrapper[4696]: I0321 08:50:09.495087 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2bbaeade-9a94-419e-8a55-3b90334ea861","Type":"ContainerDied","Data":"92865c95ba3fbb05da9e233bda7f92f1886c5f60e43902ac526e81c361e58bd6"} Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.506133 4696 generic.go:334] "Generic (PLEG): container finished" podID="e1df5e20-2002-4b15-a6b6-f23874343914" containerID="52712668b4c83dc1d0113247da7073898d3f4a1431b21f117f82c1a4d0118d02" exitCode=0 Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.506379 4696 generic.go:334] "Generic (PLEG): container finished" podID="e1df5e20-2002-4b15-a6b6-f23874343914" containerID="7eb1093e5d953b48d0bcdd3d193e4c4a05d3dd7fc4246704a7cee0195e1cb843" exitCode=0 Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.506388 4696 generic.go:334] "Generic (PLEG): container finished" podID="e1df5e20-2002-4b15-a6b6-f23874343914" containerID="bece433f574d7f51b9470e8875dfab386d4755bbb68df8962d52b36c582ba420" exitCode=0 Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.506271 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1df5e20-2002-4b15-a6b6-f23874343914","Type":"ContainerDied","Data":"52712668b4c83dc1d0113247da7073898d3f4a1431b21f117f82c1a4d0118d02"} Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.506416 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1df5e20-2002-4b15-a6b6-f23874343914","Type":"ContainerDied","Data":"7eb1093e5d953b48d0bcdd3d193e4c4a05d3dd7fc4246704a7cee0195e1cb843"} Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.506425 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1df5e20-2002-4b15-a6b6-f23874343914","Type":"ContainerDied","Data":"bece433f574d7f51b9470e8875dfab386d4755bbb68df8962d52b36c582ba420"} Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.739440 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.836295 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1df5e20-2002-4b15-a6b6-f23874343914-log-httpd\") pod \"e1df5e20-2002-4b15-a6b6-f23874343914\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.836331 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1df5e20-2002-4b15-a6b6-f23874343914-run-httpd\") pod \"e1df5e20-2002-4b15-a6b6-f23874343914\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.836428 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-scripts\") pod \"e1df5e20-2002-4b15-a6b6-f23874343914\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.836453 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-config-data\") pod \"e1df5e20-2002-4b15-a6b6-f23874343914\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.836529 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-combined-ca-bundle\") pod \"e1df5e20-2002-4b15-a6b6-f23874343914\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.836607 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-sg-core-conf-yaml\") pod \"e1df5e20-2002-4b15-a6b6-f23874343914\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.836636 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfzld\" (UniqueName: \"kubernetes.io/projected/e1df5e20-2002-4b15-a6b6-f23874343914-kube-api-access-bfzld\") pod \"e1df5e20-2002-4b15-a6b6-f23874343914\" (UID: \"e1df5e20-2002-4b15-a6b6-f23874343914\") " Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.836676 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1df5e20-2002-4b15-a6b6-f23874343914-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e1df5e20-2002-4b15-a6b6-f23874343914" (UID: "e1df5e20-2002-4b15-a6b6-f23874343914"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.836762 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1df5e20-2002-4b15-a6b6-f23874343914-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e1df5e20-2002-4b15-a6b6-f23874343914" (UID: "e1df5e20-2002-4b15-a6b6-f23874343914"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.837073 4696 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1df5e20-2002-4b15-a6b6-f23874343914-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.837086 4696 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1df5e20-2002-4b15-a6b6-f23874343914-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.847293 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-scripts" (OuterVolumeSpecName: "scripts") pod "e1df5e20-2002-4b15-a6b6-f23874343914" (UID: "e1df5e20-2002-4b15-a6b6-f23874343914"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.852055 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1df5e20-2002-4b15-a6b6-f23874343914-kube-api-access-bfzld" (OuterVolumeSpecName: "kube-api-access-bfzld") pod "e1df5e20-2002-4b15-a6b6-f23874343914" (UID: "e1df5e20-2002-4b15-a6b6-f23874343914"). InnerVolumeSpecName "kube-api-access-bfzld". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.878933 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e1df5e20-2002-4b15-a6b6-f23874343914" (UID: "e1df5e20-2002-4b15-a6b6-f23874343914"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.938501 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.938536 4696 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:10 crc kubenswrapper[4696]: I0321 08:50:10.938546 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfzld\" (UniqueName: \"kubernetes.io/projected/e1df5e20-2002-4b15-a6b6-f23874343914-kube-api-access-bfzld\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.019068 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1df5e20-2002-4b15-a6b6-f23874343914" (UID: "e1df5e20-2002-4b15-a6b6-f23874343914"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.040592 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.100610 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-config-data" (OuterVolumeSpecName: "config-data") pod "e1df5e20-2002-4b15-a6b6-f23874343914" (UID: "e1df5e20-2002-4b15-a6b6-f23874343914"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.143073 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1df5e20-2002-4b15-a6b6-f23874343914-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.518605 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1df5e20-2002-4b15-a6b6-f23874343914","Type":"ContainerDied","Data":"327aa93255173331b392ba86f6b96ae35c2656bab40ba43f69c5ae070dd5c299"} Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.518656 4696 scope.go:117] "RemoveContainer" containerID="52712668b4c83dc1d0113247da7073898d3f4a1431b21f117f82c1a4d0118d02" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.518673 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.577474 4696 scope.go:117] "RemoveContainer" containerID="7fbc416c207de813688566084124df0e65f347c6de903ae1dcdf3d7a33fdb3e4" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.619017 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.632468 4696 scope.go:117] "RemoveContainer" containerID="7eb1093e5d953b48d0bcdd3d193e4c4a05d3dd7fc4246704a7cee0195e1cb843" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.640886 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.649621 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:50:11 crc kubenswrapper[4696]: E0321 08:50:11.650148 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1df5e20-2002-4b15-a6b6-f23874343914" containerName="sg-core" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.650161 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1df5e20-2002-4b15-a6b6-f23874343914" containerName="sg-core" Mar 21 08:50:11 crc kubenswrapper[4696]: E0321 08:50:11.650171 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1df5e20-2002-4b15-a6b6-f23874343914" containerName="ceilometer-central-agent" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.650177 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1df5e20-2002-4b15-a6b6-f23874343914" containerName="ceilometer-central-agent" Mar 21 08:50:11 crc kubenswrapper[4696]: E0321 08:50:11.650207 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e8563c8-4723-4e92-9972-3cddb4c529e7" containerName="oc" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.650213 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e8563c8-4723-4e92-9972-3cddb4c529e7" containerName="oc" Mar 21 08:50:11 crc kubenswrapper[4696]: E0321 08:50:11.650227 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1df5e20-2002-4b15-a6b6-f23874343914" containerName="ceilometer-notification-agent" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.650233 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1df5e20-2002-4b15-a6b6-f23874343914" containerName="ceilometer-notification-agent" Mar 21 08:50:11 crc kubenswrapper[4696]: E0321 08:50:11.650243 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1df5e20-2002-4b15-a6b6-f23874343914" containerName="proxy-httpd" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.650248 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1df5e20-2002-4b15-a6b6-f23874343914" containerName="proxy-httpd" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.650427 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1df5e20-2002-4b15-a6b6-f23874343914" containerName="proxy-httpd" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.650450 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1df5e20-2002-4b15-a6b6-f23874343914" containerName="ceilometer-notification-agent" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.650459 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e8563c8-4723-4e92-9972-3cddb4c529e7" containerName="oc" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.650466 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1df5e20-2002-4b15-a6b6-f23874343914" containerName="sg-core" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.650478 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1df5e20-2002-4b15-a6b6-f23874343914" containerName="ceilometer-central-agent" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.652412 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.654214 4696 scope.go:117] "RemoveContainer" containerID="bece433f574d7f51b9470e8875dfab386d4755bbb68df8962d52b36c582ba420" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.654594 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.658038 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.658407 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.769383 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb1c2132-f9e1-4cae-9d03-121797e74007-log-httpd\") pod \"ceilometer-0\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.769438 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.769481 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-scripts\") pod \"ceilometer-0\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.769552 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sght9\" (UniqueName: \"kubernetes.io/projected/bb1c2132-f9e1-4cae-9d03-121797e74007-kube-api-access-sght9\") pod \"ceilometer-0\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.769596 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb1c2132-f9e1-4cae-9d03-121797e74007-run-httpd\") pod \"ceilometer-0\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.769678 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-config-data\") pod \"ceilometer-0\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.770006 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.849019 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:50:11 crc kubenswrapper[4696]: E0321 08:50:11.849779 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-sght9 log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="bb1c2132-f9e1-4cae-9d03-121797e74007" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.871837 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.871905 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb1c2132-f9e1-4cae-9d03-121797e74007-log-httpd\") pod \"ceilometer-0\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.871940 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.871992 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-scripts\") pod \"ceilometer-0\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.872036 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sght9\" (UniqueName: \"kubernetes.io/projected/bb1c2132-f9e1-4cae-9d03-121797e74007-kube-api-access-sght9\") pod \"ceilometer-0\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.872080 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb1c2132-f9e1-4cae-9d03-121797e74007-run-httpd\") pod \"ceilometer-0\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.872109 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-config-data\") pod \"ceilometer-0\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.872337 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb1c2132-f9e1-4cae-9d03-121797e74007-log-httpd\") pod \"ceilometer-0\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.872484 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb1c2132-f9e1-4cae-9d03-121797e74007-run-httpd\") pod \"ceilometer-0\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.877150 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.877858 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-scripts\") pod \"ceilometer-0\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.878417 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-config-data\") pod \"ceilometer-0\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.878876 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " pod="openstack/ceilometer-0" Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.896085 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.896309 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="c1b37b82-bd6a-4e10-98cd-89fbff6bb554" containerName="kube-state-metrics" containerID="cri-o://0660ce817029e70387eb1fe7ee9ba9f46d06dc0936e55f77f894e07fabf086af" gracePeriod=30 Mar 21 08:50:11 crc kubenswrapper[4696]: I0321 08:50:11.901149 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sght9\" (UniqueName: \"kubernetes.io/projected/bb1c2132-f9e1-4cae-9d03-121797e74007-kube-api-access-sght9\") pod \"ceilometer-0\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " pod="openstack/ceilometer-0" Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.545809 4696 generic.go:334] "Generic (PLEG): container finished" podID="c1b37b82-bd6a-4e10-98cd-89fbff6bb554" containerID="0660ce817029e70387eb1fe7ee9ba9f46d06dc0936e55f77f894e07fabf086af" exitCode=2 Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.548679 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1df5e20-2002-4b15-a6b6-f23874343914" path="/var/lib/kubelet/pods/e1df5e20-2002-4b15-a6b6-f23874343914/volumes" Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.549413 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c1b37b82-bd6a-4e10-98cd-89fbff6bb554","Type":"ContainerDied","Data":"0660ce817029e70387eb1fe7ee9ba9f46d06dc0936e55f77f894e07fabf086af"} Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.550107 4696 generic.go:334] "Generic (PLEG): container finished" podID="2bbaeade-9a94-419e-8a55-3b90334ea861" containerID="b38d446de9a50a7f1fa4339e2e05e5dcafd2d9b76fa6f8217ea05a80efc71616" exitCode=0 Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.550161 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.550229 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2bbaeade-9a94-419e-8a55-3b90334ea861","Type":"ContainerDied","Data":"b38d446de9a50a7f1fa4339e2e05e5dcafd2d9b76fa6f8217ea05a80efc71616"} Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.592984 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.695073 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-sg-core-conf-yaml\") pod \"bb1c2132-f9e1-4cae-9d03-121797e74007\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.695136 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-scripts\") pod \"bb1c2132-f9e1-4cae-9d03-121797e74007\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.695184 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-config-data\") pod \"bb1c2132-f9e1-4cae-9d03-121797e74007\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.695213 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb1c2132-f9e1-4cae-9d03-121797e74007-log-httpd\") pod \"bb1c2132-f9e1-4cae-9d03-121797e74007\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.695245 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sght9\" (UniqueName: \"kubernetes.io/projected/bb1c2132-f9e1-4cae-9d03-121797e74007-kube-api-access-sght9\") pod \"bb1c2132-f9e1-4cae-9d03-121797e74007\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.695345 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-combined-ca-bundle\") pod \"bb1c2132-f9e1-4cae-9d03-121797e74007\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.695490 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb1c2132-f9e1-4cae-9d03-121797e74007-run-httpd\") pod \"bb1c2132-f9e1-4cae-9d03-121797e74007\" (UID: \"bb1c2132-f9e1-4cae-9d03-121797e74007\") " Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.696214 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb1c2132-f9e1-4cae-9d03-121797e74007-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bb1c2132-f9e1-4cae-9d03-121797e74007" (UID: "bb1c2132-f9e1-4cae-9d03-121797e74007"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.699092 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb1c2132-f9e1-4cae-9d03-121797e74007-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bb1c2132-f9e1-4cae-9d03-121797e74007" (UID: "bb1c2132-f9e1-4cae-9d03-121797e74007"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.709002 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-scripts" (OuterVolumeSpecName: "scripts") pod "bb1c2132-f9e1-4cae-9d03-121797e74007" (UID: "bb1c2132-f9e1-4cae-9d03-121797e74007"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.709076 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bb1c2132-f9e1-4cae-9d03-121797e74007" (UID: "bb1c2132-f9e1-4cae-9d03-121797e74007"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.710372 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb1c2132-f9e1-4cae-9d03-121797e74007" (UID: "bb1c2132-f9e1-4cae-9d03-121797e74007"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.723423 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb1c2132-f9e1-4cae-9d03-121797e74007-kube-api-access-sght9" (OuterVolumeSpecName: "kube-api-access-sght9") pod "bb1c2132-f9e1-4cae-9d03-121797e74007" (UID: "bb1c2132-f9e1-4cae-9d03-121797e74007"). InnerVolumeSpecName "kube-api-access-sght9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.725953 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-config-data" (OuterVolumeSpecName: "config-data") pod "bb1c2132-f9e1-4cae-9d03-121797e74007" (UID: "bb1c2132-f9e1-4cae-9d03-121797e74007"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.799659 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.800022 4696 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb1c2132-f9e1-4cae-9d03-121797e74007-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.800034 4696 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.800047 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.800058 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb1c2132-f9e1-4cae-9d03-121797e74007-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.800070 4696 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb1c2132-f9e1-4cae-9d03-121797e74007-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:12 crc kubenswrapper[4696]: I0321 08:50:12.800080 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sght9\" (UniqueName: \"kubernetes.io/projected/bb1c2132-f9e1-4cae-9d03-121797e74007-kube-api-access-sght9\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.016295 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.105101 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dl2g\" (UniqueName: \"kubernetes.io/projected/c1b37b82-bd6a-4e10-98cd-89fbff6bb554-kube-api-access-5dl2g\") pod \"c1b37b82-bd6a-4e10-98cd-89fbff6bb554\" (UID: \"c1b37b82-bd6a-4e10-98cd-89fbff6bb554\") " Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.109503 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1b37b82-bd6a-4e10-98cd-89fbff6bb554-kube-api-access-5dl2g" (OuterVolumeSpecName: "kube-api-access-5dl2g") pod "c1b37b82-bd6a-4e10-98cd-89fbff6bb554" (UID: "c1b37b82-bd6a-4e10-98cd-89fbff6bb554"). InnerVolumeSpecName "kube-api-access-5dl2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.196350 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.207442 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dl2g\" (UniqueName: \"kubernetes.io/projected/c1b37b82-bd6a-4e10-98cd-89fbff6bb554-kube-api-access-5dl2g\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.308403 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tplb\" (UniqueName: \"kubernetes.io/projected/2bbaeade-9a94-419e-8a55-3b90334ea861-kube-api-access-8tplb\") pod \"2bbaeade-9a94-419e-8a55-3b90334ea861\" (UID: \"2bbaeade-9a94-419e-8a55-3b90334ea861\") " Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.308829 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bbaeade-9a94-419e-8a55-3b90334ea861-config-data\") pod \"2bbaeade-9a94-419e-8a55-3b90334ea861\" (UID: \"2bbaeade-9a94-419e-8a55-3b90334ea861\") " Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.308968 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bbaeade-9a94-419e-8a55-3b90334ea861-logs\") pod \"2bbaeade-9a94-419e-8a55-3b90334ea861\" (UID: \"2bbaeade-9a94-419e-8a55-3b90334ea861\") " Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.309083 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bbaeade-9a94-419e-8a55-3b90334ea861-combined-ca-bundle\") pod \"2bbaeade-9a94-419e-8a55-3b90334ea861\" (UID: \"2bbaeade-9a94-419e-8a55-3b90334ea861\") " Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.310943 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bbaeade-9a94-419e-8a55-3b90334ea861-logs" (OuterVolumeSpecName: "logs") pod "2bbaeade-9a94-419e-8a55-3b90334ea861" (UID: "2bbaeade-9a94-419e-8a55-3b90334ea861"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.315489 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bbaeade-9a94-419e-8a55-3b90334ea861-kube-api-access-8tplb" (OuterVolumeSpecName: "kube-api-access-8tplb") pod "2bbaeade-9a94-419e-8a55-3b90334ea861" (UID: "2bbaeade-9a94-419e-8a55-3b90334ea861"). InnerVolumeSpecName "kube-api-access-8tplb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.347631 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bbaeade-9a94-419e-8a55-3b90334ea861-config-data" (OuterVolumeSpecName: "config-data") pod "2bbaeade-9a94-419e-8a55-3b90334ea861" (UID: "2bbaeade-9a94-419e-8a55-3b90334ea861"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.360518 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bbaeade-9a94-419e-8a55-3b90334ea861-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2bbaeade-9a94-419e-8a55-3b90334ea861" (UID: "2bbaeade-9a94-419e-8a55-3b90334ea861"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.412027 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tplb\" (UniqueName: \"kubernetes.io/projected/2bbaeade-9a94-419e-8a55-3b90334ea861-kube-api-access-8tplb\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.412269 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bbaeade-9a94-419e-8a55-3b90334ea861-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.412346 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bbaeade-9a94-419e-8a55-3b90334ea861-logs\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.412418 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bbaeade-9a94-419e-8a55-3b90334ea861-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.559967 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c1b37b82-bd6a-4e10-98cd-89fbff6bb554","Type":"ContainerDied","Data":"4fb701a4fed069904e046a153906aef0ff0e084a70050acd1db25ff37cc27178"} Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.560014 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.560021 4696 scope.go:117] "RemoveContainer" containerID="0660ce817029e70387eb1fe7ee9ba9f46d06dc0936e55f77f894e07fabf086af" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.562194 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2bbaeade-9a94-419e-8a55-3b90334ea861","Type":"ContainerDied","Data":"a78d61dbd40752f1465b0a38d1c32371c57aa9fa2177c45af9226ee4e300c5d9"} Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.562222 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.562238 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.595479 4696 scope.go:117] "RemoveContainer" containerID="b38d446de9a50a7f1fa4339e2e05e5dcafd2d9b76fa6f8217ea05a80efc71616" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.598725 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.633066 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.640062 4696 scope.go:117] "RemoveContainer" containerID="92865c95ba3fbb05da9e233bda7f92f1886c5f60e43902ac526e81c361e58bd6" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.654353 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.679049 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 21 08:50:13 crc kubenswrapper[4696]: E0321 08:50:13.681264 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1b37b82-bd6a-4e10-98cd-89fbff6bb554" containerName="kube-state-metrics" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.681285 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1b37b82-bd6a-4e10-98cd-89fbff6bb554" containerName="kube-state-metrics" Mar 21 08:50:13 crc kubenswrapper[4696]: E0321 08:50:13.681314 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bbaeade-9a94-419e-8a55-3b90334ea861" containerName="nova-api-log" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.681320 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bbaeade-9a94-419e-8a55-3b90334ea861" containerName="nova-api-log" Mar 21 08:50:13 crc kubenswrapper[4696]: E0321 08:50:13.681351 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bbaeade-9a94-419e-8a55-3b90334ea861" containerName="nova-api-api" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.681358 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bbaeade-9a94-419e-8a55-3b90334ea861" containerName="nova-api-api" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.689208 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bbaeade-9a94-419e-8a55-3b90334ea861" containerName="nova-api-log" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.689300 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1b37b82-bd6a-4e10-98cd-89fbff6bb554" containerName="kube-state-metrics" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.689317 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bbaeade-9a94-419e-8a55-3b90334ea861" containerName="nova-api-api" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.697483 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.706125 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.706340 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-mdcrk" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.706541 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.719550 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.725034 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.747710 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.751552 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.777483 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.779371 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.789335 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.789636 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.789768 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.800102 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.822513 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53c78963-a90b-40f0-b223-0c70917cee6e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"53c78963-a90b-40f0-b223-0c70917cee6e\") " pod="openstack/kube-state-metrics-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.822644 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqnls\" (UniqueName: \"kubernetes.io/projected/53c78963-a90b-40f0-b223-0c70917cee6e-kube-api-access-kqnls\") pod \"kube-state-metrics-0\" (UID: \"53c78963-a90b-40f0-b223-0c70917cee6e\") " pod="openstack/kube-state-metrics-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.822709 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/53c78963-a90b-40f0-b223-0c70917cee6e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"53c78963-a90b-40f0-b223-0c70917cee6e\") " pod="openstack/kube-state-metrics-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.822729 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/53c78963-a90b-40f0-b223-0c70917cee6e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"53c78963-a90b-40f0-b223-0c70917cee6e\") " pod="openstack/kube-state-metrics-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.825128 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.835735 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.845873 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.848508 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.850464 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.850638 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.856233 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.924648 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6dsq\" (UniqueName: \"kubernetes.io/projected/4d4871aa-34d7-4f64-bd12-95d1b1985596-kube-api-access-d6dsq\") pod \"nova-api-0\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " pod="openstack/nova-api-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.924695 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-config-data\") pod \"ceilometer-0\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " pod="openstack/ceilometer-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.924736 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53c78963-a90b-40f0-b223-0c70917cee6e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"53c78963-a90b-40f0-b223-0c70917cee6e\") " pod="openstack/kube-state-metrics-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.924757 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-run-httpd\") pod \"ceilometer-0\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " pod="openstack/ceilometer-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.924827 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " pod="openstack/ceilometer-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.924845 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " pod="openstack/nova-api-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.924863 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " pod="openstack/nova-api-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.924894 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-log-httpd\") pod \"ceilometer-0\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " pod="openstack/ceilometer-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.924916 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqnls\" (UniqueName: \"kubernetes.io/projected/53c78963-a90b-40f0-b223-0c70917cee6e-kube-api-access-kqnls\") pod \"kube-state-metrics-0\" (UID: \"53c78963-a90b-40f0-b223-0c70917cee6e\") " pod="openstack/kube-state-metrics-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.924948 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-config-data\") pod \"nova-api-0\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " pod="openstack/nova-api-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.924966 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-scripts\") pod \"ceilometer-0\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " pod="openstack/ceilometer-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.924980 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " pod="openstack/ceilometer-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.925004 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-public-tls-certs\") pod \"nova-api-0\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " pod="openstack/nova-api-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.925024 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/53c78963-a90b-40f0-b223-0c70917cee6e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"53c78963-a90b-40f0-b223-0c70917cee6e\") " pod="openstack/kube-state-metrics-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.925044 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/53c78963-a90b-40f0-b223-0c70917cee6e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"53c78963-a90b-40f0-b223-0c70917cee6e\") " pod="openstack/kube-state-metrics-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.925083 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d4871aa-34d7-4f64-bd12-95d1b1985596-logs\") pod \"nova-api-0\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " pod="openstack/nova-api-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.925099 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncx9d\" (UniqueName: \"kubernetes.io/projected/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-kube-api-access-ncx9d\") pod \"ceilometer-0\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " pod="openstack/ceilometer-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.929766 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/53c78963-a90b-40f0-b223-0c70917cee6e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"53c78963-a90b-40f0-b223-0c70917cee6e\") " pod="openstack/kube-state-metrics-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.930505 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53c78963-a90b-40f0-b223-0c70917cee6e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"53c78963-a90b-40f0-b223-0c70917cee6e\") " pod="openstack/kube-state-metrics-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.938361 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/53c78963-a90b-40f0-b223-0c70917cee6e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"53c78963-a90b-40f0-b223-0c70917cee6e\") " pod="openstack/kube-state-metrics-0" Mar 21 08:50:13 crc kubenswrapper[4696]: I0321 08:50:13.940952 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqnls\" (UniqueName: \"kubernetes.io/projected/53c78963-a90b-40f0-b223-0c70917cee6e-kube-api-access-kqnls\") pod \"kube-state-metrics-0\" (UID: \"53c78963-a90b-40f0-b223-0c70917cee6e\") " pod="openstack/kube-state-metrics-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.027490 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " pod="openstack/ceilometer-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.027656 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " pod="openstack/nova-api-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.027691 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " pod="openstack/nova-api-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.027970 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-log-httpd\") pod \"ceilometer-0\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " pod="openstack/ceilometer-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.028084 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-config-data\") pod \"nova-api-0\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " pod="openstack/nova-api-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.028105 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-scripts\") pod \"ceilometer-0\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " pod="openstack/ceilometer-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.028123 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " pod="openstack/ceilometer-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.028512 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-log-httpd\") pod \"ceilometer-0\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " pod="openstack/ceilometer-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.028609 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-public-tls-certs\") pod \"nova-api-0\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " pod="openstack/nova-api-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.029338 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d4871aa-34d7-4f64-bd12-95d1b1985596-logs\") pod \"nova-api-0\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " pod="openstack/nova-api-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.029375 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncx9d\" (UniqueName: \"kubernetes.io/projected/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-kube-api-access-ncx9d\") pod \"ceilometer-0\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " pod="openstack/ceilometer-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.029440 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6dsq\" (UniqueName: \"kubernetes.io/projected/4d4871aa-34d7-4f64-bd12-95d1b1985596-kube-api-access-d6dsq\") pod \"nova-api-0\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " pod="openstack/nova-api-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.029465 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-config-data\") pod \"ceilometer-0\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " pod="openstack/ceilometer-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.029509 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-run-httpd\") pod \"ceilometer-0\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " pod="openstack/ceilometer-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.029776 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-run-httpd\") pod \"ceilometer-0\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " pod="openstack/ceilometer-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.029194 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.030098 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d4871aa-34d7-4f64-bd12-95d1b1985596-logs\") pod \"nova-api-0\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " pod="openstack/nova-api-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.032651 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " pod="openstack/ceilometer-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.035002 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-config-data\") pod \"ceilometer-0\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " pod="openstack/ceilometer-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.042391 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-public-tls-certs\") pod \"nova-api-0\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " pod="openstack/nova-api-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.042639 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-config-data\") pod \"nova-api-0\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " pod="openstack/nova-api-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.045077 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " pod="openstack/nova-api-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.046944 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-scripts\") pod \"ceilometer-0\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " pod="openstack/ceilometer-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.051366 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6dsq\" (UniqueName: \"kubernetes.io/projected/4d4871aa-34d7-4f64-bd12-95d1b1985596-kube-api-access-d6dsq\") pod \"nova-api-0\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " pod="openstack/nova-api-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.057085 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " pod="openstack/ceilometer-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.059579 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " pod="openstack/nova-api-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.060698 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncx9d\" (UniqueName: \"kubernetes.io/projected/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-kube-api-access-ncx9d\") pod \"ceilometer-0\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " pod="openstack/ceilometer-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.105306 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.166644 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.553163 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bbaeade-9a94-419e-8a55-3b90334ea861" path="/var/lib/kubelet/pods/2bbaeade-9a94-419e-8a55-3b90334ea861/volumes" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.554017 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb1c2132-f9e1-4cae-9d03-121797e74007" path="/var/lib/kubelet/pods/bb1c2132-f9e1-4cae-9d03-121797e74007/volumes" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.554342 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1b37b82-bd6a-4e10-98cd-89fbff6bb554" path="/var/lib/kubelet/pods/c1b37b82-bd6a-4e10-98cd-89fbff6bb554/volumes" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.610426 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.725940 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.859797 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-vzfrk"] Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.861153 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-vzfrk" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.870205 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.870600 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.887379 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-vzfrk"] Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.947663 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmwvl\" (UniqueName: \"kubernetes.io/projected/507db733-1a22-4cb0-8b84-63746b69ea4f-kube-api-access-gmwvl\") pod \"nova-cell1-cell-mapping-vzfrk\" (UID: \"507db733-1a22-4cb0-8b84-63746b69ea4f\") " pod="openstack/nova-cell1-cell-mapping-vzfrk" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.947947 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/507db733-1a22-4cb0-8b84-63746b69ea4f-config-data\") pod \"nova-cell1-cell-mapping-vzfrk\" (UID: \"507db733-1a22-4cb0-8b84-63746b69ea4f\") " pod="openstack/nova-cell1-cell-mapping-vzfrk" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.948033 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/507db733-1a22-4cb0-8b84-63746b69ea4f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-vzfrk\" (UID: \"507db733-1a22-4cb0-8b84-63746b69ea4f\") " pod="openstack/nova-cell1-cell-mapping-vzfrk" Mar 21 08:50:14 crc kubenswrapper[4696]: I0321 08:50:14.948086 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/507db733-1a22-4cb0-8b84-63746b69ea4f-scripts\") pod \"nova-cell1-cell-mapping-vzfrk\" (UID: \"507db733-1a22-4cb0-8b84-63746b69ea4f\") " pod="openstack/nova-cell1-cell-mapping-vzfrk" Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.049893 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/507db733-1a22-4cb0-8b84-63746b69ea4f-scripts\") pod \"nova-cell1-cell-mapping-vzfrk\" (UID: \"507db733-1a22-4cb0-8b84-63746b69ea4f\") " pod="openstack/nova-cell1-cell-mapping-vzfrk" Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.050151 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmwvl\" (UniqueName: \"kubernetes.io/projected/507db733-1a22-4cb0-8b84-63746b69ea4f-kube-api-access-gmwvl\") pod \"nova-cell1-cell-mapping-vzfrk\" (UID: \"507db733-1a22-4cb0-8b84-63746b69ea4f\") " pod="openstack/nova-cell1-cell-mapping-vzfrk" Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.050251 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/507db733-1a22-4cb0-8b84-63746b69ea4f-config-data\") pod \"nova-cell1-cell-mapping-vzfrk\" (UID: \"507db733-1a22-4cb0-8b84-63746b69ea4f\") " pod="openstack/nova-cell1-cell-mapping-vzfrk" Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.050277 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/507db733-1a22-4cb0-8b84-63746b69ea4f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-vzfrk\" (UID: \"507db733-1a22-4cb0-8b84-63746b69ea4f\") " pod="openstack/nova-cell1-cell-mapping-vzfrk" Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.053854 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/507db733-1a22-4cb0-8b84-63746b69ea4f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-vzfrk\" (UID: \"507db733-1a22-4cb0-8b84-63746b69ea4f\") " pod="openstack/nova-cell1-cell-mapping-vzfrk" Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.057138 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/507db733-1a22-4cb0-8b84-63746b69ea4f-scripts\") pod \"nova-cell1-cell-mapping-vzfrk\" (UID: \"507db733-1a22-4cb0-8b84-63746b69ea4f\") " pod="openstack/nova-cell1-cell-mapping-vzfrk" Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.060752 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/507db733-1a22-4cb0-8b84-63746b69ea4f-config-data\") pod \"nova-cell1-cell-mapping-vzfrk\" (UID: \"507db733-1a22-4cb0-8b84-63746b69ea4f\") " pod="openstack/nova-cell1-cell-mapping-vzfrk" Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.061829 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.073777 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmwvl\" (UniqueName: \"kubernetes.io/projected/507db733-1a22-4cb0-8b84-63746b69ea4f-kube-api-access-gmwvl\") pod \"nova-cell1-cell-mapping-vzfrk\" (UID: \"507db733-1a22-4cb0-8b84-63746b69ea4f\") " pod="openstack/nova-cell1-cell-mapping-vzfrk" Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.131687 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.189657 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-vzfrk" Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.215796 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.616788 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f2c69dd-df7e-4693-af3c-2994e40ef3c6","Type":"ContainerStarted","Data":"818ecef01db75a14ad9f354aa0bc89b32e9dd700205bdc49e60d5fd7b62a7aa9"} Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.626711 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4d4871aa-34d7-4f64-bd12-95d1b1985596","Type":"ContainerStarted","Data":"b7b0df2f77f3bfdda9e92a98ded5e0f71bbad817a34143e9f481264f25297d82"} Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.626757 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4d4871aa-34d7-4f64-bd12-95d1b1985596","Type":"ContainerStarted","Data":"b55a03f1af07a0c2de3c6029c2db2075f9f639ecb51ee30c3681888dcd572d0b"} Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.639344 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"53c78963-a90b-40f0-b223-0c70917cee6e","Type":"ContainerStarted","Data":"6ab429acb81c07025cf0ab5539f386024d4e710d92a8f2a3488f39f420456aea"} Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.639406 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"53c78963-a90b-40f0-b223-0c70917cee6e","Type":"ContainerStarted","Data":"99a4607bd93875c42eec02d0bf006eede0bd07ff92f1ef6986fda800e56e8406"} Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.639550 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.757526 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.3322722430000002 podStartE2EDuration="2.757505069s" podCreationTimestamp="2026-03-21 08:50:13 +0000 UTC" firstStartedPulling="2026-03-21 08:50:14.617403297 +0000 UTC m=+1348.738284010" lastFinishedPulling="2026-03-21 08:50:15.042636123 +0000 UTC m=+1349.163516836" observedRunningTime="2026-03-21 08:50:15.670293263 +0000 UTC m=+1349.791173976" watchObservedRunningTime="2026-03-21 08:50:15.757505069 +0000 UTC m=+1349.878385782" Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.763655 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-vzfrk"] Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.921949 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.992172 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78cd565959-cg8dx"] Mar 21 08:50:15 crc kubenswrapper[4696]: I0321 08:50:15.992422 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-78cd565959-cg8dx" podUID="3f429da6-5739-4444-8b85-347dc971df0f" containerName="dnsmasq-dns" containerID="cri-o://7cb8c4ea50917367d7fc916a934a75bbe68e00c5926ac97cdd8289edf3d76842" gracePeriod=10 Mar 21 08:50:16 crc kubenswrapper[4696]: I0321 08:50:16.678212 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4d4871aa-34d7-4f64-bd12-95d1b1985596","Type":"ContainerStarted","Data":"5b7d89bcc92a03af4000a4ae44361b68e5dc059859ddb3234f69d36d7675ca46"} Mar 21 08:50:16 crc kubenswrapper[4696]: I0321 08:50:16.686212 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-vzfrk" event={"ID":"507db733-1a22-4cb0-8b84-63746b69ea4f","Type":"ContainerStarted","Data":"9886082c780636b6eb0d3b13b235a0763fe4ce7830032888f8c170f2e7893b1c"} Mar 21 08:50:16 crc kubenswrapper[4696]: I0321 08:50:16.686252 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-vzfrk" event={"ID":"507db733-1a22-4cb0-8b84-63746b69ea4f","Type":"ContainerStarted","Data":"8769692617ffdd1671534bacd063fbb7a86599f8698c561a7e031147c272e987"} Mar 21 08:50:16 crc kubenswrapper[4696]: I0321 08:50:16.694140 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f2c69dd-df7e-4693-af3c-2994e40ef3c6","Type":"ContainerStarted","Data":"8be7e70b5ab55e0b915b6f69b25325b134535ec898b8bee86b05a6f2448c600a"} Mar 21 08:50:16 crc kubenswrapper[4696]: I0321 08:50:16.700715 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.700697387 podStartE2EDuration="3.700697387s" podCreationTimestamp="2026-03-21 08:50:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:50:16.69608498 +0000 UTC m=+1350.816965693" watchObservedRunningTime="2026-03-21 08:50:16.700697387 +0000 UTC m=+1350.821578100" Mar 21 08:50:16 crc kubenswrapper[4696]: I0321 08:50:16.711704 4696 generic.go:334] "Generic (PLEG): container finished" podID="3f429da6-5739-4444-8b85-347dc971df0f" containerID="7cb8c4ea50917367d7fc916a934a75bbe68e00c5926ac97cdd8289edf3d76842" exitCode=0 Mar 21 08:50:16 crc kubenswrapper[4696]: I0321 08:50:16.711764 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78cd565959-cg8dx" event={"ID":"3f429da6-5739-4444-8b85-347dc971df0f","Type":"ContainerDied","Data":"7cb8c4ea50917367d7fc916a934a75bbe68e00c5926ac97cdd8289edf3d76842"} Mar 21 08:50:16 crc kubenswrapper[4696]: I0321 08:50:16.720759 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-vzfrk" podStartSLOduration=2.720735303 podStartE2EDuration="2.720735303s" podCreationTimestamp="2026-03-21 08:50:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:50:16.716068613 +0000 UTC m=+1350.836949346" watchObservedRunningTime="2026-03-21 08:50:16.720735303 +0000 UTC m=+1350.841616016" Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.039582 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.118392 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-dns-swift-storage-0\") pod \"3f429da6-5739-4444-8b85-347dc971df0f\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.118442 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-dns-svc\") pod \"3f429da6-5739-4444-8b85-347dc971df0f\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.118482 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-config\") pod \"3f429da6-5739-4444-8b85-347dc971df0f\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.118596 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9dxq\" (UniqueName: \"kubernetes.io/projected/3f429da6-5739-4444-8b85-347dc971df0f-kube-api-access-l9dxq\") pod \"3f429da6-5739-4444-8b85-347dc971df0f\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.118702 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-ovsdbserver-sb\") pod \"3f429da6-5739-4444-8b85-347dc971df0f\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.118772 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-ovsdbserver-nb\") pod \"3f429da6-5739-4444-8b85-347dc971df0f\" (UID: \"3f429da6-5739-4444-8b85-347dc971df0f\") " Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.155576 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f429da6-5739-4444-8b85-347dc971df0f-kube-api-access-l9dxq" (OuterVolumeSpecName: "kube-api-access-l9dxq") pod "3f429da6-5739-4444-8b85-347dc971df0f" (UID: "3f429da6-5739-4444-8b85-347dc971df0f"). InnerVolumeSpecName "kube-api-access-l9dxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.212350 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3f429da6-5739-4444-8b85-347dc971df0f" (UID: "3f429da6-5739-4444-8b85-347dc971df0f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.221915 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3f429da6-5739-4444-8b85-347dc971df0f" (UID: "3f429da6-5739-4444-8b85-347dc971df0f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.223372 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.223401 4696 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.223412 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9dxq\" (UniqueName: \"kubernetes.io/projected/3f429da6-5739-4444-8b85-347dc971df0f-kube-api-access-l9dxq\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.243861 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3f429da6-5739-4444-8b85-347dc971df0f" (UID: "3f429da6-5739-4444-8b85-347dc971df0f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.258241 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-config" (OuterVolumeSpecName: "config") pod "3f429da6-5739-4444-8b85-347dc971df0f" (UID: "3f429da6-5739-4444-8b85-347dc971df0f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.280734 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3f429da6-5739-4444-8b85-347dc971df0f" (UID: "3f429da6-5739-4444-8b85-347dc971df0f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.325284 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.325576 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.325589 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f429da6-5739-4444-8b85-347dc971df0f-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.724534 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f2c69dd-df7e-4693-af3c-2994e40ef3c6","Type":"ContainerStarted","Data":"c3aef397831812a583dd42e2ef0e1312c6bce143e3dd83fb261d7a08c14a1bb3"} Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.724599 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f2c69dd-df7e-4693-af3c-2994e40ef3c6","Type":"ContainerStarted","Data":"9b01bd5aa6fc3f20dab47626e790a91860bb9d513fefca06d36d05932efb29cc"} Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.726608 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78cd565959-cg8dx" event={"ID":"3f429da6-5739-4444-8b85-347dc971df0f","Type":"ContainerDied","Data":"3adf528c84939fa36d51ea254f38bb9553a43b017ea883c60b489b02f5eb9e96"} Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.726704 4696 scope.go:117] "RemoveContainer" containerID="7cb8c4ea50917367d7fc916a934a75bbe68e00c5926ac97cdd8289edf3d76842" Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.726723 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78cd565959-cg8dx" Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.747762 4696 scope.go:117] "RemoveContainer" containerID="37130251119ee5a8480000c6dc37701adcac361a80e5141bae0b5b2c6e4fcf13" Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.758637 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78cd565959-cg8dx"] Mar 21 08:50:17 crc kubenswrapper[4696]: I0321 08:50:17.768253 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78cd565959-cg8dx"] Mar 21 08:50:18 crc kubenswrapper[4696]: I0321 08:50:18.546865 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f429da6-5739-4444-8b85-347dc971df0f" path="/var/lib/kubelet/pods/3f429da6-5739-4444-8b85-347dc971df0f/volumes" Mar 21 08:50:20 crc kubenswrapper[4696]: I0321 08:50:20.792805 4696 generic.go:334] "Generic (PLEG): container finished" podID="0f2c69dd-df7e-4693-af3c-2994e40ef3c6" containerID="debb172dbbb0689dc117174fdc8fc3a2a1a59c22d1ed2c4a341a608b12983534" exitCode=1 Mar 21 08:50:20 crc kubenswrapper[4696]: I0321 08:50:20.792862 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f2c69dd-df7e-4693-af3c-2994e40ef3c6","Type":"ContainerDied","Data":"debb172dbbb0689dc117174fdc8fc3a2a1a59c22d1ed2c4a341a608b12983534"} Mar 21 08:50:20 crc kubenswrapper[4696]: I0321 08:50:20.793011 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0f2c69dd-df7e-4693-af3c-2994e40ef3c6" containerName="ceilometer-central-agent" containerID="cri-o://8be7e70b5ab55e0b915b6f69b25325b134535ec898b8bee86b05a6f2448c600a" gracePeriod=30 Mar 21 08:50:20 crc kubenswrapper[4696]: I0321 08:50:20.793041 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0f2c69dd-df7e-4693-af3c-2994e40ef3c6" containerName="ceilometer-notification-agent" containerID="cri-o://c3aef397831812a583dd42e2ef0e1312c6bce143e3dd83fb261d7a08c14a1bb3" gracePeriod=30 Mar 21 08:50:20 crc kubenswrapper[4696]: I0321 08:50:20.793073 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0f2c69dd-df7e-4693-af3c-2994e40ef3c6" containerName="sg-core" containerID="cri-o://9b01bd5aa6fc3f20dab47626e790a91860bb9d513fefca06d36d05932efb29cc" gracePeriod=30 Mar 21 08:50:21 crc kubenswrapper[4696]: I0321 08:50:21.805209 4696 generic.go:334] "Generic (PLEG): container finished" podID="507db733-1a22-4cb0-8b84-63746b69ea4f" containerID="9886082c780636b6eb0d3b13b235a0763fe4ce7830032888f8c170f2e7893b1c" exitCode=0 Mar 21 08:50:21 crc kubenswrapper[4696]: I0321 08:50:21.805293 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-vzfrk" event={"ID":"507db733-1a22-4cb0-8b84-63746b69ea4f","Type":"ContainerDied","Data":"9886082c780636b6eb0d3b13b235a0763fe4ce7830032888f8c170f2e7893b1c"} Mar 21 08:50:21 crc kubenswrapper[4696]: I0321 08:50:21.808686 4696 generic.go:334] "Generic (PLEG): container finished" podID="0f2c69dd-df7e-4693-af3c-2994e40ef3c6" containerID="9b01bd5aa6fc3f20dab47626e790a91860bb9d513fefca06d36d05932efb29cc" exitCode=2 Mar 21 08:50:21 crc kubenswrapper[4696]: I0321 08:50:21.808711 4696 generic.go:334] "Generic (PLEG): container finished" podID="0f2c69dd-df7e-4693-af3c-2994e40ef3c6" containerID="c3aef397831812a583dd42e2ef0e1312c6bce143e3dd83fb261d7a08c14a1bb3" exitCode=0 Mar 21 08:50:21 crc kubenswrapper[4696]: I0321 08:50:21.808727 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f2c69dd-df7e-4693-af3c-2994e40ef3c6","Type":"ContainerDied","Data":"9b01bd5aa6fc3f20dab47626e790a91860bb9d513fefca06d36d05932efb29cc"} Mar 21 08:50:21 crc kubenswrapper[4696]: I0321 08:50:21.808756 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f2c69dd-df7e-4693-af3c-2994e40ef3c6","Type":"ContainerDied","Data":"c3aef397831812a583dd42e2ef0e1312c6bce143e3dd83fb261d7a08c14a1bb3"} Mar 21 08:50:22 crc kubenswrapper[4696]: I0321 08:50:22.822582 4696 generic.go:334] "Generic (PLEG): container finished" podID="0f2c69dd-df7e-4693-af3c-2994e40ef3c6" containerID="8be7e70b5ab55e0b915b6f69b25325b134535ec898b8bee86b05a6f2448c600a" exitCode=0 Mar 21 08:50:22 crc kubenswrapper[4696]: I0321 08:50:22.822653 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f2c69dd-df7e-4693-af3c-2994e40ef3c6","Type":"ContainerDied","Data":"8be7e70b5ab55e0b915b6f69b25325b134535ec898b8bee86b05a6f2448c600a"} Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.733284 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.738113 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-vzfrk" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.834182 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-vzfrk" event={"ID":"507db733-1a22-4cb0-8b84-63746b69ea4f","Type":"ContainerDied","Data":"8769692617ffdd1671534bacd063fbb7a86599f8698c561a7e031147c272e987"} Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.834210 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-vzfrk" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.834221 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8769692617ffdd1671534bacd063fbb7a86599f8698c561a7e031147c272e987" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.843553 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f2c69dd-df7e-4693-af3c-2994e40ef3c6","Type":"ContainerDied","Data":"818ecef01db75a14ad9f354aa0bc89b32e9dd700205bdc49e60d5fd7b62a7aa9"} Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.843605 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.843616 4696 scope.go:117] "RemoveContainer" containerID="debb172dbbb0689dc117174fdc8fc3a2a1a59c22d1ed2c4a341a608b12983534" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.849917 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-sg-core-conf-yaml\") pod \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.850001 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-config-data\") pod \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.850053 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-run-httpd\") pod \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.850091 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncx9d\" (UniqueName: \"kubernetes.io/projected/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-kube-api-access-ncx9d\") pod \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.850127 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-log-httpd\") pod \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.850223 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-combined-ca-bundle\") pod \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.850253 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-scripts\") pod \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\" (UID: \"0f2c69dd-df7e-4693-af3c-2994e40ef3c6\") " Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.850290 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/507db733-1a22-4cb0-8b84-63746b69ea4f-scripts\") pod \"507db733-1a22-4cb0-8b84-63746b69ea4f\" (UID: \"507db733-1a22-4cb0-8b84-63746b69ea4f\") " Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.850376 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/507db733-1a22-4cb0-8b84-63746b69ea4f-config-data\") pod \"507db733-1a22-4cb0-8b84-63746b69ea4f\" (UID: \"507db733-1a22-4cb0-8b84-63746b69ea4f\") " Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.850408 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmwvl\" (UniqueName: \"kubernetes.io/projected/507db733-1a22-4cb0-8b84-63746b69ea4f-kube-api-access-gmwvl\") pod \"507db733-1a22-4cb0-8b84-63746b69ea4f\" (UID: \"507db733-1a22-4cb0-8b84-63746b69ea4f\") " Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.850431 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/507db733-1a22-4cb0-8b84-63746b69ea4f-combined-ca-bundle\") pod \"507db733-1a22-4cb0-8b84-63746b69ea4f\" (UID: \"507db733-1a22-4cb0-8b84-63746b69ea4f\") " Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.853074 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0f2c69dd-df7e-4693-af3c-2994e40ef3c6" (UID: "0f2c69dd-df7e-4693-af3c-2994e40ef3c6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.854032 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0f2c69dd-df7e-4693-af3c-2994e40ef3c6" (UID: "0f2c69dd-df7e-4693-af3c-2994e40ef3c6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.857451 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-kube-api-access-ncx9d" (OuterVolumeSpecName: "kube-api-access-ncx9d") pod "0f2c69dd-df7e-4693-af3c-2994e40ef3c6" (UID: "0f2c69dd-df7e-4693-af3c-2994e40ef3c6"). InnerVolumeSpecName "kube-api-access-ncx9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.857772 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/507db733-1a22-4cb0-8b84-63746b69ea4f-kube-api-access-gmwvl" (OuterVolumeSpecName: "kube-api-access-gmwvl") pod "507db733-1a22-4cb0-8b84-63746b69ea4f" (UID: "507db733-1a22-4cb0-8b84-63746b69ea4f"). InnerVolumeSpecName "kube-api-access-gmwvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.864398 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/507db733-1a22-4cb0-8b84-63746b69ea4f-scripts" (OuterVolumeSpecName: "scripts") pod "507db733-1a22-4cb0-8b84-63746b69ea4f" (UID: "507db733-1a22-4cb0-8b84-63746b69ea4f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.870102 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-scripts" (OuterVolumeSpecName: "scripts") pod "0f2c69dd-df7e-4693-af3c-2994e40ef3c6" (UID: "0f2c69dd-df7e-4693-af3c-2994e40ef3c6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.878174 4696 scope.go:117] "RemoveContainer" containerID="9b01bd5aa6fc3f20dab47626e790a91860bb9d513fefca06d36d05932efb29cc" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.884684 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0f2c69dd-df7e-4693-af3c-2994e40ef3c6" (UID: "0f2c69dd-df7e-4693-af3c-2994e40ef3c6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.885298 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/507db733-1a22-4cb0-8b84-63746b69ea4f-config-data" (OuterVolumeSpecName: "config-data") pod "507db733-1a22-4cb0-8b84-63746b69ea4f" (UID: "507db733-1a22-4cb0-8b84-63746b69ea4f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.885645 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/507db733-1a22-4cb0-8b84-63746b69ea4f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "507db733-1a22-4cb0-8b84-63746b69ea4f" (UID: "507db733-1a22-4cb0-8b84-63746b69ea4f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.952946 4696 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.952971 4696 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.952981 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncx9d\" (UniqueName: \"kubernetes.io/projected/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-kube-api-access-ncx9d\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.952992 4696 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.953002 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.953012 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/507db733-1a22-4cb0-8b84-63746b69ea4f-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.953022 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/507db733-1a22-4cb0-8b84-63746b69ea4f-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.953032 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmwvl\" (UniqueName: \"kubernetes.io/projected/507db733-1a22-4cb0-8b84-63746b69ea4f-kube-api-access-gmwvl\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.953041 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/507db733-1a22-4cb0-8b84-63746b69ea4f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.971996 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f2c69dd-df7e-4693-af3c-2994e40ef3c6" (UID: "0f2c69dd-df7e-4693-af3c-2994e40ef3c6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:23 crc kubenswrapper[4696]: I0321 08:50:23.983914 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-config-data" (OuterVolumeSpecName: "config-data") pod "0f2c69dd-df7e-4693-af3c-2994e40ef3c6" (UID: "0f2c69dd-df7e-4693-af3c-2994e40ef3c6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.013784 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.018261 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4d4871aa-34d7-4f64-bd12-95d1b1985596" containerName="nova-api-api" containerID="cri-o://5b7d89bcc92a03af4000a4ae44361b68e5dc059859ddb3234f69d36d7675ca46" gracePeriod=30 Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.018219 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4d4871aa-34d7-4f64-bd12-95d1b1985596" containerName="nova-api-log" containerID="cri-o://b7b0df2f77f3bfdda9e92a98ded5e0f71bbad817a34143e9f481264f25297d82" gracePeriod=30 Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.036746 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.036962 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="910aaf27-d4d5-4031-9971-95b472f065d8" containerName="nova-scheduler-scheduler" containerID="cri-o://1cae17529197bff02832d32292e43ac560bf06d48c328a2323d64f5a3646602d" gracePeriod=30 Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.054648 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.054880 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.055537 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f2c69dd-df7e-4693-af3c-2994e40ef3c6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.100728 4696 scope.go:117] "RemoveContainer" containerID="c3aef397831812a583dd42e2ef0e1312c6bce143e3dd83fb261d7a08c14a1bb3" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.124158 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.124387 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fb5d65be-1467-4ef3-9a33-c7707c4b23d2" containerName="nova-metadata-log" containerID="cri-o://bcdfb356288d9965ba7a3b0a90fc27796a87029add34df6a18564f8e4ff98898" gracePeriod=30 Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.124515 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fb5d65be-1467-4ef3-9a33-c7707c4b23d2" containerName="nova-metadata-metadata" containerID="cri-o://14fa5fb139c9f816fc8709e2ea4371101baea2bfb37b5e603c3c80787b9c3fbd" gracePeriod=30 Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.133085 4696 scope.go:117] "RemoveContainer" containerID="8be7e70b5ab55e0b915b6f69b25325b134535ec898b8bee86b05a6f2448c600a" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.241176 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.273651 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.284395 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:50:24 crc kubenswrapper[4696]: E0321 08:50:24.284926 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f2c69dd-df7e-4693-af3c-2994e40ef3c6" containerName="sg-core" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.284944 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f2c69dd-df7e-4693-af3c-2994e40ef3c6" containerName="sg-core" Mar 21 08:50:24 crc kubenswrapper[4696]: E0321 08:50:24.284960 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f2c69dd-df7e-4693-af3c-2994e40ef3c6" containerName="proxy-httpd" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.284968 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f2c69dd-df7e-4693-af3c-2994e40ef3c6" containerName="proxy-httpd" Mar 21 08:50:24 crc kubenswrapper[4696]: E0321 08:50:24.284985 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f2c69dd-df7e-4693-af3c-2994e40ef3c6" containerName="ceilometer-central-agent" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.284992 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f2c69dd-df7e-4693-af3c-2994e40ef3c6" containerName="ceilometer-central-agent" Mar 21 08:50:24 crc kubenswrapper[4696]: E0321 08:50:24.285004 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f2c69dd-df7e-4693-af3c-2994e40ef3c6" containerName="ceilometer-notification-agent" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.285012 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f2c69dd-df7e-4693-af3c-2994e40ef3c6" containerName="ceilometer-notification-agent" Mar 21 08:50:24 crc kubenswrapper[4696]: E0321 08:50:24.285027 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f429da6-5739-4444-8b85-347dc971df0f" containerName="dnsmasq-dns" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.285034 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f429da6-5739-4444-8b85-347dc971df0f" containerName="dnsmasq-dns" Mar 21 08:50:24 crc kubenswrapper[4696]: E0321 08:50:24.285051 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f429da6-5739-4444-8b85-347dc971df0f" containerName="init" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.285058 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f429da6-5739-4444-8b85-347dc971df0f" containerName="init" Mar 21 08:50:24 crc kubenswrapper[4696]: E0321 08:50:24.285080 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="507db733-1a22-4cb0-8b84-63746b69ea4f" containerName="nova-manage" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.285088 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="507db733-1a22-4cb0-8b84-63746b69ea4f" containerName="nova-manage" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.285340 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f2c69dd-df7e-4693-af3c-2994e40ef3c6" containerName="ceilometer-central-agent" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.285359 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f2c69dd-df7e-4693-af3c-2994e40ef3c6" containerName="proxy-httpd" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.285369 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="507db733-1a22-4cb0-8b84-63746b69ea4f" containerName="nova-manage" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.285381 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f2c69dd-df7e-4693-af3c-2994e40ef3c6" containerName="ceilometer-notification-agent" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.285388 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f429da6-5739-4444-8b85-347dc971df0f" containerName="dnsmasq-dns" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.285406 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f2c69dd-df7e-4693-af3c-2994e40ef3c6" containerName="sg-core" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.299118 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.299784 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.306942 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.308357 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.308599 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.362082 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-scripts\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.362435 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.362475 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-run-httpd\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.362525 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-config-data\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.362553 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.362587 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhj99\" (UniqueName: \"kubernetes.io/projected/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-kube-api-access-hhj99\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.362636 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.362676 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-log-httpd\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.465019 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-scripts\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.465130 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.465178 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-run-httpd\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.465242 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-config-data\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.465289 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.465339 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhj99\" (UniqueName: \"kubernetes.io/projected/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-kube-api-access-hhj99\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.465449 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.465772 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-run-httpd\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.465861 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-log-httpd\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.466203 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-log-httpd\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.470124 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-scripts\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.470280 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-config-data\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.476213 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.480109 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.484379 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.489572 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhj99\" (UniqueName: \"kubernetes.io/projected/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-kube-api-access-hhj99\") pod \"ceilometer-0\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.545987 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f2c69dd-df7e-4693-af3c-2994e40ef3c6" path="/var/lib/kubelet/pods/0f2c69dd-df7e-4693-af3c-2994e40ef3c6/volumes" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.620457 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.866217 4696 generic.go:334] "Generic (PLEG): container finished" podID="fb5d65be-1467-4ef3-9a33-c7707c4b23d2" containerID="bcdfb356288d9965ba7a3b0a90fc27796a87029add34df6a18564f8e4ff98898" exitCode=143 Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.866545 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb5d65be-1467-4ef3-9a33-c7707c4b23d2","Type":"ContainerDied","Data":"bcdfb356288d9965ba7a3b0a90fc27796a87029add34df6a18564f8e4ff98898"} Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.871609 4696 generic.go:334] "Generic (PLEG): container finished" podID="4d4871aa-34d7-4f64-bd12-95d1b1985596" containerID="5b7d89bcc92a03af4000a4ae44361b68e5dc059859ddb3234f69d36d7675ca46" exitCode=0 Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.871632 4696 generic.go:334] "Generic (PLEG): container finished" podID="4d4871aa-34d7-4f64-bd12-95d1b1985596" containerID="b7b0df2f77f3bfdda9e92a98ded5e0f71bbad817a34143e9f481264f25297d82" exitCode=143 Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.871652 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4d4871aa-34d7-4f64-bd12-95d1b1985596","Type":"ContainerDied","Data":"5b7d89bcc92a03af4000a4ae44361b68e5dc059859ddb3234f69d36d7675ca46"} Mar 21 08:50:24 crc kubenswrapper[4696]: I0321 08:50:24.871692 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4d4871aa-34d7-4f64-bd12-95d1b1985596","Type":"ContainerDied","Data":"b7b0df2f77f3bfdda9e92a98ded5e0f71bbad817a34143e9f481264f25297d82"} Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.066243 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.168584 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:50:25 crc kubenswrapper[4696]: W0321 08:50:25.169426 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e6b9c7f_6556_4ab0_ac07_4c11a662000a.slice/crio-f53c5d75d83f40a10774ac7a9c7a0386687b3d7627fa966f16eb8892db66ff75 WatchSource:0}: Error finding container f53c5d75d83f40a10774ac7a9c7a0386687b3d7627fa966f16eb8892db66ff75: Status 404 returned error can't find the container with id f53c5d75d83f40a10774ac7a9c7a0386687b3d7627fa966f16eb8892db66ff75 Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.195787 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-combined-ca-bundle\") pod \"4d4871aa-34d7-4f64-bd12-95d1b1985596\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.195943 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-public-tls-certs\") pod \"4d4871aa-34d7-4f64-bd12-95d1b1985596\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.195978 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-internal-tls-certs\") pod \"4d4871aa-34d7-4f64-bd12-95d1b1985596\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.196030 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6dsq\" (UniqueName: \"kubernetes.io/projected/4d4871aa-34d7-4f64-bd12-95d1b1985596-kube-api-access-d6dsq\") pod \"4d4871aa-34d7-4f64-bd12-95d1b1985596\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.196130 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-config-data\") pod \"4d4871aa-34d7-4f64-bd12-95d1b1985596\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.196297 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d4871aa-34d7-4f64-bd12-95d1b1985596-logs\") pod \"4d4871aa-34d7-4f64-bd12-95d1b1985596\" (UID: \"4d4871aa-34d7-4f64-bd12-95d1b1985596\") " Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.197184 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d4871aa-34d7-4f64-bd12-95d1b1985596-logs" (OuterVolumeSpecName: "logs") pod "4d4871aa-34d7-4f64-bd12-95d1b1985596" (UID: "4d4871aa-34d7-4f64-bd12-95d1b1985596"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.208012 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d4871aa-34d7-4f64-bd12-95d1b1985596-kube-api-access-d6dsq" (OuterVolumeSpecName: "kube-api-access-d6dsq") pod "4d4871aa-34d7-4f64-bd12-95d1b1985596" (UID: "4d4871aa-34d7-4f64-bd12-95d1b1985596"). InnerVolumeSpecName "kube-api-access-d6dsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.247906 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-config-data" (OuterVolumeSpecName: "config-data") pod "4d4871aa-34d7-4f64-bd12-95d1b1985596" (UID: "4d4871aa-34d7-4f64-bd12-95d1b1985596"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.260762 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d4871aa-34d7-4f64-bd12-95d1b1985596" (UID: "4d4871aa-34d7-4f64-bd12-95d1b1985596"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.283414 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4d4871aa-34d7-4f64-bd12-95d1b1985596" (UID: "4d4871aa-34d7-4f64-bd12-95d1b1985596"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.298358 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.298390 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d4871aa-34d7-4f64-bd12-95d1b1985596-logs\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.298413 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.298422 4696 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.298431 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6dsq\" (UniqueName: \"kubernetes.io/projected/4d4871aa-34d7-4f64-bd12-95d1b1985596-kube-api-access-d6dsq\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.300887 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4d4871aa-34d7-4f64-bd12-95d1b1985596" (UID: "4d4871aa-34d7-4f64-bd12-95d1b1985596"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.401178 4696 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d4871aa-34d7-4f64-bd12-95d1b1985596-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.900173 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e6b9c7f-6556-4ab0-ac07-4c11a662000a","Type":"ContainerStarted","Data":"9d00b013cc0ef90677d79280d7ad247cccf2daf66eaf6f11cb75fd883f410c37"} Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.900221 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e6b9c7f-6556-4ab0-ac07-4c11a662000a","Type":"ContainerStarted","Data":"f53c5d75d83f40a10774ac7a9c7a0386687b3d7627fa966f16eb8892db66ff75"} Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.922968 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4d4871aa-34d7-4f64-bd12-95d1b1985596","Type":"ContainerDied","Data":"b55a03f1af07a0c2de3c6029c2db2075f9f639ecb51ee30c3681888dcd572d0b"} Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.923017 4696 scope.go:117] "RemoveContainer" containerID="5b7d89bcc92a03af4000a4ae44361b68e5dc059859ddb3234f69d36d7675ca46" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.923179 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.961137 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.963154 4696 scope.go:117] "RemoveContainer" containerID="b7b0df2f77f3bfdda9e92a98ded5e0f71bbad817a34143e9f481264f25297d82" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.970339 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.978392 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 21 08:50:25 crc kubenswrapper[4696]: E0321 08:50:25.978798 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d4871aa-34d7-4f64-bd12-95d1b1985596" containerName="nova-api-api" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.978827 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d4871aa-34d7-4f64-bd12-95d1b1985596" containerName="nova-api-api" Mar 21 08:50:25 crc kubenswrapper[4696]: E0321 08:50:25.978849 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d4871aa-34d7-4f64-bd12-95d1b1985596" containerName="nova-api-log" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.978854 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d4871aa-34d7-4f64-bd12-95d1b1985596" containerName="nova-api-log" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.979046 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d4871aa-34d7-4f64-bd12-95d1b1985596" containerName="nova-api-api" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.979058 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d4871aa-34d7-4f64-bd12-95d1b1985596" containerName="nova-api-log" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.980135 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.982545 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.982978 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.983462 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 21 08:50:25 crc kubenswrapper[4696]: I0321 08:50:25.990964 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.116229 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/348bda58-214e-41e3-bc04-2fdcf0e79142-public-tls-certs\") pod \"nova-api-0\" (UID: \"348bda58-214e-41e3-bc04-2fdcf0e79142\") " pod="openstack/nova-api-0" Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.116316 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/348bda58-214e-41e3-bc04-2fdcf0e79142-config-data\") pod \"nova-api-0\" (UID: \"348bda58-214e-41e3-bc04-2fdcf0e79142\") " pod="openstack/nova-api-0" Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.116478 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/348bda58-214e-41e3-bc04-2fdcf0e79142-logs\") pod \"nova-api-0\" (UID: \"348bda58-214e-41e3-bc04-2fdcf0e79142\") " pod="openstack/nova-api-0" Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.116638 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r46h9\" (UniqueName: \"kubernetes.io/projected/348bda58-214e-41e3-bc04-2fdcf0e79142-kube-api-access-r46h9\") pod \"nova-api-0\" (UID: \"348bda58-214e-41e3-bc04-2fdcf0e79142\") " pod="openstack/nova-api-0" Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.116761 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/348bda58-214e-41e3-bc04-2fdcf0e79142-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"348bda58-214e-41e3-bc04-2fdcf0e79142\") " pod="openstack/nova-api-0" Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.116827 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/348bda58-214e-41e3-bc04-2fdcf0e79142-internal-tls-certs\") pod \"nova-api-0\" (UID: \"348bda58-214e-41e3-bc04-2fdcf0e79142\") " pod="openstack/nova-api-0" Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.218859 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/348bda58-214e-41e3-bc04-2fdcf0e79142-config-data\") pod \"nova-api-0\" (UID: \"348bda58-214e-41e3-bc04-2fdcf0e79142\") " pod="openstack/nova-api-0" Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.218912 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/348bda58-214e-41e3-bc04-2fdcf0e79142-logs\") pod \"nova-api-0\" (UID: \"348bda58-214e-41e3-bc04-2fdcf0e79142\") " pod="openstack/nova-api-0" Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.218949 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r46h9\" (UniqueName: \"kubernetes.io/projected/348bda58-214e-41e3-bc04-2fdcf0e79142-kube-api-access-r46h9\") pod \"nova-api-0\" (UID: \"348bda58-214e-41e3-bc04-2fdcf0e79142\") " pod="openstack/nova-api-0" Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.218992 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/348bda58-214e-41e3-bc04-2fdcf0e79142-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"348bda58-214e-41e3-bc04-2fdcf0e79142\") " pod="openstack/nova-api-0" Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.219016 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/348bda58-214e-41e3-bc04-2fdcf0e79142-internal-tls-certs\") pod \"nova-api-0\" (UID: \"348bda58-214e-41e3-bc04-2fdcf0e79142\") " pod="openstack/nova-api-0" Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.219092 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/348bda58-214e-41e3-bc04-2fdcf0e79142-public-tls-certs\") pod \"nova-api-0\" (UID: \"348bda58-214e-41e3-bc04-2fdcf0e79142\") " pod="openstack/nova-api-0" Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.221218 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/348bda58-214e-41e3-bc04-2fdcf0e79142-logs\") pod \"nova-api-0\" (UID: \"348bda58-214e-41e3-bc04-2fdcf0e79142\") " pod="openstack/nova-api-0" Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.223440 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/348bda58-214e-41e3-bc04-2fdcf0e79142-config-data\") pod \"nova-api-0\" (UID: \"348bda58-214e-41e3-bc04-2fdcf0e79142\") " pod="openstack/nova-api-0" Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.223628 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/348bda58-214e-41e3-bc04-2fdcf0e79142-internal-tls-certs\") pod \"nova-api-0\" (UID: \"348bda58-214e-41e3-bc04-2fdcf0e79142\") " pod="openstack/nova-api-0" Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.224441 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/348bda58-214e-41e3-bc04-2fdcf0e79142-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"348bda58-214e-41e3-bc04-2fdcf0e79142\") " pod="openstack/nova-api-0" Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.225324 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/348bda58-214e-41e3-bc04-2fdcf0e79142-public-tls-certs\") pod \"nova-api-0\" (UID: \"348bda58-214e-41e3-bc04-2fdcf0e79142\") " pod="openstack/nova-api-0" Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.235517 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r46h9\" (UniqueName: \"kubernetes.io/projected/348bda58-214e-41e3-bc04-2fdcf0e79142-kube-api-access-r46h9\") pod \"nova-api-0\" (UID: \"348bda58-214e-41e3-bc04-2fdcf0e79142\") " pod="openstack/nova-api-0" Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.302123 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.552676 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d4871aa-34d7-4f64-bd12-95d1b1985596" path="/var/lib/kubelet/pods/4d4871aa-34d7-4f64-bd12-95d1b1985596/volumes" Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.818408 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 21 08:50:26 crc kubenswrapper[4696]: W0321 08:50:26.819634 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod348bda58_214e_41e3_bc04_2fdcf0e79142.slice/crio-709a2228d1a8b8728735316d585c101719e7a7f155b7f1e45bf1008f44d32237 WatchSource:0}: Error finding container 709a2228d1a8b8728735316d585c101719e7a7f155b7f1e45bf1008f44d32237: Status 404 returned error can't find the container with id 709a2228d1a8b8728735316d585c101719e7a7f155b7f1e45bf1008f44d32237 Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.941999 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"348bda58-214e-41e3-bc04-2fdcf0e79142","Type":"ContainerStarted","Data":"709a2228d1a8b8728735316d585c101719e7a7f155b7f1e45bf1008f44d32237"} Mar 21 08:50:26 crc kubenswrapper[4696]: I0321 08:50:26.950398 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e6b9c7f-6556-4ab0-ac07-4c11a662000a","Type":"ContainerStarted","Data":"04bd95501bb3b6bac5fa56ada1fc312ed788d6a33ec38d1182f877c9bc2eeed2"} Mar 21 08:50:27 crc kubenswrapper[4696]: E0321 08:50:27.571313 4696 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1cae17529197bff02832d32292e43ac560bf06d48c328a2323d64f5a3646602d is running failed: container process not found" containerID="1cae17529197bff02832d32292e43ac560bf06d48c328a2323d64f5a3646602d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 21 08:50:27 crc kubenswrapper[4696]: E0321 08:50:27.571902 4696 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1cae17529197bff02832d32292e43ac560bf06d48c328a2323d64f5a3646602d is running failed: container process not found" containerID="1cae17529197bff02832d32292e43ac560bf06d48c328a2323d64f5a3646602d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 21 08:50:27 crc kubenswrapper[4696]: E0321 08:50:27.575919 4696 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1cae17529197bff02832d32292e43ac560bf06d48c328a2323d64f5a3646602d is running failed: container process not found" containerID="1cae17529197bff02832d32292e43ac560bf06d48c328a2323d64f5a3646602d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 21 08:50:27 crc kubenswrapper[4696]: E0321 08:50:27.575991 4696 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1cae17529197bff02832d32292e43ac560bf06d48c328a2323d64f5a3646602d is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="910aaf27-d4d5-4031-9971-95b472f065d8" containerName="nova-scheduler-scheduler" Mar 21 08:50:27 crc kubenswrapper[4696]: I0321 08:50:27.740731 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 21 08:50:27 crc kubenswrapper[4696]: I0321 08:50:27.850125 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/910aaf27-d4d5-4031-9971-95b472f065d8-config-data\") pod \"910aaf27-d4d5-4031-9971-95b472f065d8\" (UID: \"910aaf27-d4d5-4031-9971-95b472f065d8\") " Mar 21 08:50:27 crc kubenswrapper[4696]: I0321 08:50:27.850176 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4748t\" (UniqueName: \"kubernetes.io/projected/910aaf27-d4d5-4031-9971-95b472f065d8-kube-api-access-4748t\") pod \"910aaf27-d4d5-4031-9971-95b472f065d8\" (UID: \"910aaf27-d4d5-4031-9971-95b472f065d8\") " Mar 21 08:50:27 crc kubenswrapper[4696]: I0321 08:50:27.850294 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/910aaf27-d4d5-4031-9971-95b472f065d8-combined-ca-bundle\") pod \"910aaf27-d4d5-4031-9971-95b472f065d8\" (UID: \"910aaf27-d4d5-4031-9971-95b472f065d8\") " Mar 21 08:50:27 crc kubenswrapper[4696]: I0321 08:50:27.867150 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/910aaf27-d4d5-4031-9971-95b472f065d8-kube-api-access-4748t" (OuterVolumeSpecName: "kube-api-access-4748t") pod "910aaf27-d4d5-4031-9971-95b472f065d8" (UID: "910aaf27-d4d5-4031-9971-95b472f065d8"). InnerVolumeSpecName "kube-api-access-4748t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:50:27 crc kubenswrapper[4696]: I0321 08:50:27.890386 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/910aaf27-d4d5-4031-9971-95b472f065d8-config-data" (OuterVolumeSpecName: "config-data") pod "910aaf27-d4d5-4031-9971-95b472f065d8" (UID: "910aaf27-d4d5-4031-9971-95b472f065d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:27 crc kubenswrapper[4696]: I0321 08:50:27.909491 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/910aaf27-d4d5-4031-9971-95b472f065d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "910aaf27-d4d5-4031-9971-95b472f065d8" (UID: "910aaf27-d4d5-4031-9971-95b472f065d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:27 crc kubenswrapper[4696]: I0321 08:50:27.952027 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/910aaf27-d4d5-4031-9971-95b472f065d8-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:27 crc kubenswrapper[4696]: I0321 08:50:27.952056 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4748t\" (UniqueName: \"kubernetes.io/projected/910aaf27-d4d5-4031-9971-95b472f065d8-kube-api-access-4748t\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:27 crc kubenswrapper[4696]: I0321 08:50:27.952069 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/910aaf27-d4d5-4031-9971-95b472f065d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:27 crc kubenswrapper[4696]: I0321 08:50:27.972190 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"348bda58-214e-41e3-bc04-2fdcf0e79142","Type":"ContainerStarted","Data":"c59d0a591a0948bb0f591b57ceacbccd543440e9939a88ae399198270813b38d"} Mar 21 08:50:27 crc kubenswrapper[4696]: I0321 08:50:27.972256 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"348bda58-214e-41e3-bc04-2fdcf0e79142","Type":"ContainerStarted","Data":"002aa5e1285a14bc7460e792c698babdb12679f0f3a705ea397baef3bac78a65"} Mar 21 08:50:27 crc kubenswrapper[4696]: I0321 08:50:27.986496 4696 generic.go:334] "Generic (PLEG): container finished" podID="910aaf27-d4d5-4031-9971-95b472f065d8" containerID="1cae17529197bff02832d32292e43ac560bf06d48c328a2323d64f5a3646602d" exitCode=0 Mar 21 08:50:27 crc kubenswrapper[4696]: I0321 08:50:27.986550 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"910aaf27-d4d5-4031-9971-95b472f065d8","Type":"ContainerDied","Data":"1cae17529197bff02832d32292e43ac560bf06d48c328a2323d64f5a3646602d"} Mar 21 08:50:27 crc kubenswrapper[4696]: I0321 08:50:27.986576 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"910aaf27-d4d5-4031-9971-95b472f065d8","Type":"ContainerDied","Data":"19e6f747772077f00d1425247e124bf2530bdaf1860b5cee3cbd1a69eeed7529"} Mar 21 08:50:27 crc kubenswrapper[4696]: I0321 08:50:27.986592 4696 scope.go:117] "RemoveContainer" containerID="1cae17529197bff02832d32292e43ac560bf06d48c328a2323d64f5a3646602d" Mar 21 08:50:27 crc kubenswrapper[4696]: I0321 08:50:27.986744 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.008904 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e6b9c7f-6556-4ab0-ac07-4c11a662000a","Type":"ContainerStarted","Data":"c3316c4e2ea7b7880570f92fd7258613ef9e26551661cba16cd26594b8e33910"} Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.011754 4696 generic.go:334] "Generic (PLEG): container finished" podID="fb5d65be-1467-4ef3-9a33-c7707c4b23d2" containerID="14fa5fb139c9f816fc8709e2ea4371101baea2bfb37b5e603c3c80787b9c3fbd" exitCode=0 Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.011779 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb5d65be-1467-4ef3-9a33-c7707c4b23d2","Type":"ContainerDied","Data":"14fa5fb139c9f816fc8709e2ea4371101baea2bfb37b5e603c3c80787b9c3fbd"} Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.019153 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.019133708 podStartE2EDuration="3.019133708s" podCreationTimestamp="2026-03-21 08:50:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:50:28.003692664 +0000 UTC m=+1362.124573377" watchObservedRunningTime="2026-03-21 08:50:28.019133708 +0000 UTC m=+1362.140014421" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.044227 4696 scope.go:117] "RemoveContainer" containerID="1cae17529197bff02832d32292e43ac560bf06d48c328a2323d64f5a3646602d" Mar 21 08:50:28 crc kubenswrapper[4696]: E0321 08:50:28.064970 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cae17529197bff02832d32292e43ac560bf06d48c328a2323d64f5a3646602d\": container with ID starting with 1cae17529197bff02832d32292e43ac560bf06d48c328a2323d64f5a3646602d not found: ID does not exist" containerID="1cae17529197bff02832d32292e43ac560bf06d48c328a2323d64f5a3646602d" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.065025 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cae17529197bff02832d32292e43ac560bf06d48c328a2323d64f5a3646602d"} err="failed to get container status \"1cae17529197bff02832d32292e43ac560bf06d48c328a2323d64f5a3646602d\": rpc error: code = NotFound desc = could not find container \"1cae17529197bff02832d32292e43ac560bf06d48c328a2323d64f5a3646602d\": container with ID starting with 1cae17529197bff02832d32292e43ac560bf06d48c328a2323d64f5a3646602d not found: ID does not exist" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.078884 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.107962 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.116554 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 08:50:28 crc kubenswrapper[4696]: E0321 08:50:28.117081 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="910aaf27-d4d5-4031-9971-95b472f065d8" containerName="nova-scheduler-scheduler" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.117095 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="910aaf27-d4d5-4031-9971-95b472f065d8" containerName="nova-scheduler-scheduler" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.117311 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="910aaf27-d4d5-4031-9971-95b472f065d8" containerName="nova-scheduler-scheduler" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.118586 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.124312 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.126738 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.167118 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b1800dd-28c2-4614-be8a-06afc002a85e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8b1800dd-28c2-4614-be8a-06afc002a85e\") " pod="openstack/nova-scheduler-0" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.167614 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmq5l\" (UniqueName: \"kubernetes.io/projected/8b1800dd-28c2-4614-be8a-06afc002a85e-kube-api-access-gmq5l\") pod \"nova-scheduler-0\" (UID: \"8b1800dd-28c2-4614-be8a-06afc002a85e\") " pod="openstack/nova-scheduler-0" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.167743 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b1800dd-28c2-4614-be8a-06afc002a85e-config-data\") pod \"nova-scheduler-0\" (UID: \"8b1800dd-28c2-4614-be8a-06afc002a85e\") " pod="openstack/nova-scheduler-0" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.271063 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b1800dd-28c2-4614-be8a-06afc002a85e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8b1800dd-28c2-4614-be8a-06afc002a85e\") " pod="openstack/nova-scheduler-0" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.271407 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmq5l\" (UniqueName: \"kubernetes.io/projected/8b1800dd-28c2-4614-be8a-06afc002a85e-kube-api-access-gmq5l\") pod \"nova-scheduler-0\" (UID: \"8b1800dd-28c2-4614-be8a-06afc002a85e\") " pod="openstack/nova-scheduler-0" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.271429 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b1800dd-28c2-4614-be8a-06afc002a85e-config-data\") pod \"nova-scheduler-0\" (UID: \"8b1800dd-28c2-4614-be8a-06afc002a85e\") " pod="openstack/nova-scheduler-0" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.275629 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b1800dd-28c2-4614-be8a-06afc002a85e-config-data\") pod \"nova-scheduler-0\" (UID: \"8b1800dd-28c2-4614-be8a-06afc002a85e\") " pod="openstack/nova-scheduler-0" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.276358 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b1800dd-28c2-4614-be8a-06afc002a85e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8b1800dd-28c2-4614-be8a-06afc002a85e\") " pod="openstack/nova-scheduler-0" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.290836 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmq5l\" (UniqueName: \"kubernetes.io/projected/8b1800dd-28c2-4614-be8a-06afc002a85e-kube-api-access-gmq5l\") pod \"nova-scheduler-0\" (UID: \"8b1800dd-28c2-4614-be8a-06afc002a85e\") " pod="openstack/nova-scheduler-0" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.357581 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.472993 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.473737 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-logs\") pod \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\" (UID: \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\") " Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.473768 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-config-data\") pod \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\" (UID: \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\") " Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.474096 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-nova-metadata-tls-certs\") pod \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\" (UID: \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\") " Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.474168 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdxq6\" (UniqueName: \"kubernetes.io/projected/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-kube-api-access-sdxq6\") pod \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\" (UID: \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\") " Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.474233 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-combined-ca-bundle\") pod \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\" (UID: \"fb5d65be-1467-4ef3-9a33-c7707c4b23d2\") " Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.483154 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-kube-api-access-sdxq6" (OuterVolumeSpecName: "kube-api-access-sdxq6") pod "fb5d65be-1467-4ef3-9a33-c7707c4b23d2" (UID: "fb5d65be-1467-4ef3-9a33-c7707c4b23d2"). InnerVolumeSpecName "kube-api-access-sdxq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.486745 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-logs" (OuterVolumeSpecName: "logs") pod "fb5d65be-1467-4ef3-9a33-c7707c4b23d2" (UID: "fb5d65be-1467-4ef3-9a33-c7707c4b23d2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.511579 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-config-data" (OuterVolumeSpecName: "config-data") pod "fb5d65be-1467-4ef3-9a33-c7707c4b23d2" (UID: "fb5d65be-1467-4ef3-9a33-c7707c4b23d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.533793 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb5d65be-1467-4ef3-9a33-c7707c4b23d2" (UID: "fb5d65be-1467-4ef3-9a33-c7707c4b23d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.556739 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="910aaf27-d4d5-4031-9971-95b472f065d8" path="/var/lib/kubelet/pods/910aaf27-d4d5-4031-9971-95b472f065d8/volumes" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.562854 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "fb5d65be-1467-4ef3-9a33-c7707c4b23d2" (UID: "fb5d65be-1467-4ef3-9a33-c7707c4b23d2"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.579344 4696 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.579391 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdxq6\" (UniqueName: \"kubernetes.io/projected/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-kube-api-access-sdxq6\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.579402 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.579411 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-logs\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:28 crc kubenswrapper[4696]: I0321 08:50:28.579420 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb5d65be-1467-4ef3-9a33-c7707c4b23d2-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.030673 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.031557 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb5d65be-1467-4ef3-9a33-c7707c4b23d2","Type":"ContainerDied","Data":"bba16a65a33465580d3dfe23a3df54d414630960638aa0c79ced29884712d231"} Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.031596 4696 scope.go:117] "RemoveContainer" containerID="14fa5fb139c9f816fc8709e2ea4371101baea2bfb37b5e603c3c80787b9c3fbd" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.213877 4696 scope.go:117] "RemoveContainer" containerID="bcdfb356288d9965ba7a3b0a90fc27796a87029add34df6a18564f8e4ff98898" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.248090 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.271964 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.291219 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 21 08:50:29 crc kubenswrapper[4696]: E0321 08:50:29.291630 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb5d65be-1467-4ef3-9a33-c7707c4b23d2" containerName="nova-metadata-metadata" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.291648 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb5d65be-1467-4ef3-9a33-c7707c4b23d2" containerName="nova-metadata-metadata" Mar 21 08:50:29 crc kubenswrapper[4696]: E0321 08:50:29.291677 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb5d65be-1467-4ef3-9a33-c7707c4b23d2" containerName="nova-metadata-log" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.291684 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb5d65be-1467-4ef3-9a33-c7707c4b23d2" containerName="nova-metadata-log" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.300535 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb5d65be-1467-4ef3-9a33-c7707c4b23d2" containerName="nova-metadata-log" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.300574 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb5d65be-1467-4ef3-9a33-c7707c4b23d2" containerName="nova-metadata-metadata" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.302534 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.306269 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.306455 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.315494 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 08:50:29 crc kubenswrapper[4696]: W0321 08:50:29.351121 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b1800dd_28c2_4614_be8a_06afc002a85e.slice/crio-5c76e66bcdd3123e9e79c0de0d8c47c9c3b6c504ed31da38b300dfc60f6c1d65 WatchSource:0}: Error finding container 5c76e66bcdd3123e9e79c0de0d8c47c9c3b6c504ed31da38b300dfc60f6c1d65: Status 404 returned error can't find the container with id 5c76e66bcdd3123e9e79c0de0d8c47c9c3b6c504ed31da38b300dfc60f6c1d65 Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.352858 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.399202 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f812f449-0a51-4414-ac6f-31ecf7648686-logs\") pod \"nova-metadata-0\" (UID: \"f812f449-0a51-4414-ac6f-31ecf7648686\") " pod="openstack/nova-metadata-0" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.399603 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f812f449-0a51-4414-ac6f-31ecf7648686-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f812f449-0a51-4414-ac6f-31ecf7648686\") " pod="openstack/nova-metadata-0" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.399713 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f812f449-0a51-4414-ac6f-31ecf7648686-config-data\") pod \"nova-metadata-0\" (UID: \"f812f449-0a51-4414-ac6f-31ecf7648686\") " pod="openstack/nova-metadata-0" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.400010 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f812f449-0a51-4414-ac6f-31ecf7648686-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f812f449-0a51-4414-ac6f-31ecf7648686\") " pod="openstack/nova-metadata-0" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.400124 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh8f4\" (UniqueName: \"kubernetes.io/projected/f812f449-0a51-4414-ac6f-31ecf7648686-kube-api-access-qh8f4\") pod \"nova-metadata-0\" (UID: \"f812f449-0a51-4414-ac6f-31ecf7648686\") " pod="openstack/nova-metadata-0" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.501383 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f812f449-0a51-4414-ac6f-31ecf7648686-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f812f449-0a51-4414-ac6f-31ecf7648686\") " pod="openstack/nova-metadata-0" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.501639 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f812f449-0a51-4414-ac6f-31ecf7648686-config-data\") pod \"nova-metadata-0\" (UID: \"f812f449-0a51-4414-ac6f-31ecf7648686\") " pod="openstack/nova-metadata-0" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.501911 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f812f449-0a51-4414-ac6f-31ecf7648686-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f812f449-0a51-4414-ac6f-31ecf7648686\") " pod="openstack/nova-metadata-0" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.502118 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh8f4\" (UniqueName: \"kubernetes.io/projected/f812f449-0a51-4414-ac6f-31ecf7648686-kube-api-access-qh8f4\") pod \"nova-metadata-0\" (UID: \"f812f449-0a51-4414-ac6f-31ecf7648686\") " pod="openstack/nova-metadata-0" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.502280 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f812f449-0a51-4414-ac6f-31ecf7648686-logs\") pod \"nova-metadata-0\" (UID: \"f812f449-0a51-4414-ac6f-31ecf7648686\") " pod="openstack/nova-metadata-0" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.502715 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f812f449-0a51-4414-ac6f-31ecf7648686-logs\") pod \"nova-metadata-0\" (UID: \"f812f449-0a51-4414-ac6f-31ecf7648686\") " pod="openstack/nova-metadata-0" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.505172 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f812f449-0a51-4414-ac6f-31ecf7648686-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f812f449-0a51-4414-ac6f-31ecf7648686\") " pod="openstack/nova-metadata-0" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.505747 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f812f449-0a51-4414-ac6f-31ecf7648686-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f812f449-0a51-4414-ac6f-31ecf7648686\") " pod="openstack/nova-metadata-0" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.506181 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f812f449-0a51-4414-ac6f-31ecf7648686-config-data\") pod \"nova-metadata-0\" (UID: \"f812f449-0a51-4414-ac6f-31ecf7648686\") " pod="openstack/nova-metadata-0" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.522463 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh8f4\" (UniqueName: \"kubernetes.io/projected/f812f449-0a51-4414-ac6f-31ecf7648686-kube-api-access-qh8f4\") pod \"nova-metadata-0\" (UID: \"f812f449-0a51-4414-ac6f-31ecf7648686\") " pod="openstack/nova-metadata-0" Mar 21 08:50:29 crc kubenswrapper[4696]: I0321 08:50:29.626695 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 21 08:50:30 crc kubenswrapper[4696]: I0321 08:50:30.041923 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8b1800dd-28c2-4614-be8a-06afc002a85e","Type":"ContainerStarted","Data":"74c49b057631bb716dbfc8c87c921f91465f2dca771317b7eddb9859fb6b871e"} Mar 21 08:50:30 crc kubenswrapper[4696]: I0321 08:50:30.042317 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8b1800dd-28c2-4614-be8a-06afc002a85e","Type":"ContainerStarted","Data":"5c76e66bcdd3123e9e79c0de0d8c47c9c3b6c504ed31da38b300dfc60f6c1d65"} Mar 21 08:50:30 crc kubenswrapper[4696]: I0321 08:50:30.048625 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e6b9c7f-6556-4ab0-ac07-4c11a662000a","Type":"ContainerStarted","Data":"f84e937cd55f41687d35f38f8a0e1f5b0d2a53c681b57c74d51a161e910e9ed5"} Mar 21 08:50:30 crc kubenswrapper[4696]: I0321 08:50:30.048790 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 21 08:50:30 crc kubenswrapper[4696]: I0321 08:50:30.082588 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.082566072 podStartE2EDuration="2.082566072s" podCreationTimestamp="2026-03-21 08:50:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:50:30.05721426 +0000 UTC m=+1364.178094993" watchObservedRunningTime="2026-03-21 08:50:30.082566072 +0000 UTC m=+1364.203446785" Mar 21 08:50:30 crc kubenswrapper[4696]: I0321 08:50:30.102005 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.275673934 podStartE2EDuration="6.101973117s" podCreationTimestamp="2026-03-21 08:50:24 +0000 UTC" firstStartedPulling="2026-03-21 08:50:25.173100682 +0000 UTC m=+1359.293981395" lastFinishedPulling="2026-03-21 08:50:28.999399865 +0000 UTC m=+1363.120280578" observedRunningTime="2026-03-21 08:50:30.076889883 +0000 UTC m=+1364.197770616" watchObservedRunningTime="2026-03-21 08:50:30.101973117 +0000 UTC m=+1364.222853830" Mar 21 08:50:30 crc kubenswrapper[4696]: I0321 08:50:30.147129 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 21 08:50:30 crc kubenswrapper[4696]: I0321 08:50:30.341240 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:50:30 crc kubenswrapper[4696]: I0321 08:50:30.341563 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:50:30 crc kubenswrapper[4696]: I0321 08:50:30.550832 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb5d65be-1467-4ef3-9a33-c7707c4b23d2" path="/var/lib/kubelet/pods/fb5d65be-1467-4ef3-9a33-c7707c4b23d2/volumes" Mar 21 08:50:31 crc kubenswrapper[4696]: I0321 08:50:31.061514 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f812f449-0a51-4414-ac6f-31ecf7648686","Type":"ContainerStarted","Data":"31f3e52196848de8df9feff1e7218a45c03ac3b235b9aa44e106879bccc861e2"} Mar 21 08:50:31 crc kubenswrapper[4696]: I0321 08:50:31.061807 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f812f449-0a51-4414-ac6f-31ecf7648686","Type":"ContainerStarted","Data":"344d70871dc08512a894dc60bf3601cd092bac764802174f4b8fa981d542ddbe"} Mar 21 08:50:31 crc kubenswrapper[4696]: I0321 08:50:31.061845 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f812f449-0a51-4414-ac6f-31ecf7648686","Type":"ContainerStarted","Data":"ac1d302ab7341f30a721c2383f0e67e8a311b300431669e4a278bcc3f689c205"} Mar 21 08:50:31 crc kubenswrapper[4696]: I0321 08:50:31.082950 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.082935283 podStartE2EDuration="2.082935283s" podCreationTimestamp="2026-03-21 08:50:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:50:31.07893095 +0000 UTC m=+1365.199811663" watchObservedRunningTime="2026-03-21 08:50:31.082935283 +0000 UTC m=+1365.203815996" Mar 21 08:50:33 crc kubenswrapper[4696]: I0321 08:50:33.474222 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 21 08:50:36 crc kubenswrapper[4696]: I0321 08:50:36.303045 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 21 08:50:36 crc kubenswrapper[4696]: I0321 08:50:36.303644 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 21 08:50:37 crc kubenswrapper[4696]: I0321 08:50:37.317991 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="348bda58-214e-41e3-bc04-2fdcf0e79142" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.239:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 21 08:50:37 crc kubenswrapper[4696]: I0321 08:50:37.318972 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="348bda58-214e-41e3-bc04-2fdcf0e79142" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.239:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 21 08:50:38 crc kubenswrapper[4696]: I0321 08:50:38.474347 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 21 08:50:38 crc kubenswrapper[4696]: I0321 08:50:38.504250 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 21 08:50:39 crc kubenswrapper[4696]: I0321 08:50:39.162980 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 21 08:50:39 crc kubenswrapper[4696]: I0321 08:50:39.627195 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 21 08:50:39 crc kubenswrapper[4696]: I0321 08:50:39.627252 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 21 08:50:40 crc kubenswrapper[4696]: I0321 08:50:40.641076 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f812f449-0a51-4414-ac6f-31ecf7648686" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.241:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 21 08:50:40 crc kubenswrapper[4696]: I0321 08:50:40.641308 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f812f449-0a51-4414-ac6f-31ecf7648686" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.241:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 21 08:50:44 crc kubenswrapper[4696]: I0321 08:50:44.302661 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 21 08:50:44 crc kubenswrapper[4696]: I0321 08:50:44.303085 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 21 08:50:46 crc kubenswrapper[4696]: I0321 08:50:46.308946 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 21 08:50:46 crc kubenswrapper[4696]: I0321 08:50:46.313641 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 21 08:50:46 crc kubenswrapper[4696]: I0321 08:50:46.316262 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 21 08:50:47 crc kubenswrapper[4696]: I0321 08:50:47.235629 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 21 08:50:47 crc kubenswrapper[4696]: I0321 08:50:47.627716 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 21 08:50:47 crc kubenswrapper[4696]: I0321 08:50:47.628799 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 21 08:50:49 crc kubenswrapper[4696]: I0321 08:50:49.632163 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 21 08:50:49 crc kubenswrapper[4696]: I0321 08:50:49.634996 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 21 08:50:49 crc kubenswrapper[4696]: I0321 08:50:49.639700 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 21 08:50:50 crc kubenswrapper[4696]: I0321 08:50:50.264148 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 21 08:50:54 crc kubenswrapper[4696]: I0321 08:50:54.629715 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 21 08:50:57 crc kubenswrapper[4696]: I0321 08:50:57.250382 4696 scope.go:117] "RemoveContainer" containerID="25b39e1721f34b8f1fc2ea8ab54b44f3e42a606c54d75eec4d1ef997119951b3" Mar 21 08:51:00 crc kubenswrapper[4696]: I0321 08:51:00.341010 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:51:00 crc kubenswrapper[4696]: I0321 08:51:00.341280 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.323783 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-db-sync-vgsz6"] Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.332806 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-db-sync-vgsz6"] Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.426183 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-db-sync-fq9xb"] Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.427607 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-fq9xb" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.435069 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.455530 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-fq9xb"] Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.546598 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-config-data\") pod \"cloudkitty-db-sync-fq9xb\" (UID: \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\") " pod="openstack/cloudkitty-db-sync-fq9xb" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.546731 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-certs\") pod \"cloudkitty-db-sync-fq9xb\" (UID: \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\") " pod="openstack/cloudkitty-db-sync-fq9xb" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.546786 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kzqv\" (UniqueName: \"kubernetes.io/projected/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-kube-api-access-5kzqv\") pod \"cloudkitty-db-sync-fq9xb\" (UID: \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\") " pod="openstack/cloudkitty-db-sync-fq9xb" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.546858 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-combined-ca-bundle\") pod \"cloudkitty-db-sync-fq9xb\" (UID: \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\") " pod="openstack/cloudkitty-db-sync-fq9xb" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.546931 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-scripts\") pod \"cloudkitty-db-sync-fq9xb\" (UID: \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\") " pod="openstack/cloudkitty-db-sync-fq9xb" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.615301 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3fc5362-53e5-4c35-be23-49dc3dd1fcba" path="/var/lib/kubelet/pods/d3fc5362-53e5-4c35-be23-49dc3dd1fcba/volumes" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.630883 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-br9c4"] Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.632994 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-br9c4" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.648578 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-config-data\") pod \"cloudkitty-db-sync-fq9xb\" (UID: \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\") " pod="openstack/cloudkitty-db-sync-fq9xb" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.648660 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-certs\") pod \"cloudkitty-db-sync-fq9xb\" (UID: \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\") " pod="openstack/cloudkitty-db-sync-fq9xb" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.648709 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kzqv\" (UniqueName: \"kubernetes.io/projected/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-kube-api-access-5kzqv\") pod \"cloudkitty-db-sync-fq9xb\" (UID: \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\") " pod="openstack/cloudkitty-db-sync-fq9xb" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.648733 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-combined-ca-bundle\") pod \"cloudkitty-db-sync-fq9xb\" (UID: \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\") " pod="openstack/cloudkitty-db-sync-fq9xb" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.648760 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-scripts\") pod \"cloudkitty-db-sync-fq9xb\" (UID: \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\") " pod="openstack/cloudkitty-db-sync-fq9xb" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.659154 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-combined-ca-bundle\") pod \"cloudkitty-db-sync-fq9xb\" (UID: \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\") " pod="openstack/cloudkitty-db-sync-fq9xb" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.661742 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-br9c4"] Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.664730 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-scripts\") pod \"cloudkitty-db-sync-fq9xb\" (UID: \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\") " pod="openstack/cloudkitty-db-sync-fq9xb" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.666746 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-config-data\") pod \"cloudkitty-db-sync-fq9xb\" (UID: \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\") " pod="openstack/cloudkitty-db-sync-fq9xb" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.686405 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kzqv\" (UniqueName: \"kubernetes.io/projected/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-kube-api-access-5kzqv\") pod \"cloudkitty-db-sync-fq9xb\" (UID: \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\") " pod="openstack/cloudkitty-db-sync-fq9xb" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.686463 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-certs\") pod \"cloudkitty-db-sync-fq9xb\" (UID: \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\") " pod="openstack/cloudkitty-db-sync-fq9xb" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.747747 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-fq9xb" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.750506 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfntv\" (UniqueName: \"kubernetes.io/projected/817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a-kube-api-access-bfntv\") pod \"redhat-operators-br9c4\" (UID: \"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a\") " pod="openshift-marketplace/redhat-operators-br9c4" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.750584 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a-catalog-content\") pod \"redhat-operators-br9c4\" (UID: \"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a\") " pod="openshift-marketplace/redhat-operators-br9c4" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.750650 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a-utilities\") pod \"redhat-operators-br9c4\" (UID: \"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a\") " pod="openshift-marketplace/redhat-operators-br9c4" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.853050 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a-catalog-content\") pod \"redhat-operators-br9c4\" (UID: \"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a\") " pod="openshift-marketplace/redhat-operators-br9c4" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.853157 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a-utilities\") pod \"redhat-operators-br9c4\" (UID: \"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a\") " pod="openshift-marketplace/redhat-operators-br9c4" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.853246 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfntv\" (UniqueName: \"kubernetes.io/projected/817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a-kube-api-access-bfntv\") pod \"redhat-operators-br9c4\" (UID: \"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a\") " pod="openshift-marketplace/redhat-operators-br9c4" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.853753 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a-catalog-content\") pod \"redhat-operators-br9c4\" (UID: \"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a\") " pod="openshift-marketplace/redhat-operators-br9c4" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.853753 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a-utilities\") pod \"redhat-operators-br9c4\" (UID: \"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a\") " pod="openshift-marketplace/redhat-operators-br9c4" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.872877 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfntv\" (UniqueName: \"kubernetes.io/projected/817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a-kube-api-access-bfntv\") pod \"redhat-operators-br9c4\" (UID: \"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a\") " pod="openshift-marketplace/redhat-operators-br9c4" Mar 21 08:51:06 crc kubenswrapper[4696]: I0321 08:51:06.955162 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-br9c4" Mar 21 08:51:07 crc kubenswrapper[4696]: I0321 08:51:07.491267 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-fq9xb"] Mar 21 08:51:07 crc kubenswrapper[4696]: I0321 08:51:07.496682 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 08:51:07 crc kubenswrapper[4696]: I0321 08:51:07.577160 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-br9c4"] Mar 21 08:51:07 crc kubenswrapper[4696]: W0321 08:51:07.581596 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod817cb85b_e1e8_4ab7_93d9_2f7e78b8fc5a.slice/crio-8b8a41951178c42396c7ecdfd19805743ec595d6378554a8f7008c7b82fc49c9 WatchSource:0}: Error finding container 8b8a41951178c42396c7ecdfd19805743ec595d6378554a8f7008c7b82fc49c9: Status 404 returned error can't find the container with id 8b8a41951178c42396c7ecdfd19805743ec595d6378554a8f7008c7b82fc49c9 Mar 21 08:51:08 crc kubenswrapper[4696]: I0321 08:51:08.333708 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 21 08:51:08 crc kubenswrapper[4696]: I0321 08:51:08.456149 4696 generic.go:334] "Generic (PLEG): container finished" podID="817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a" containerID="58e01952ce0256947c0ae7181c811bc2bd35b1555b6ba25a6b7f5d438f2726f1" exitCode=0 Mar 21 08:51:08 crc kubenswrapper[4696]: I0321 08:51:08.456215 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-br9c4" event={"ID":"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a","Type":"ContainerDied","Data":"58e01952ce0256947c0ae7181c811bc2bd35b1555b6ba25a6b7f5d438f2726f1"} Mar 21 08:51:08 crc kubenswrapper[4696]: I0321 08:51:08.456241 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-br9c4" event={"ID":"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a","Type":"ContainerStarted","Data":"8b8a41951178c42396c7ecdfd19805743ec595d6378554a8f7008c7b82fc49c9"} Mar 21 08:51:08 crc kubenswrapper[4696]: I0321 08:51:08.466301 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-fq9xb" event={"ID":"1f8a2d50-6415-4d3a-8cb4-bbbfce785842","Type":"ContainerStarted","Data":"d1d6775ac797c68b94bd80ce81230d525dd64dd919c8958f29709727cbef35a3"} Mar 21 08:51:09 crc kubenswrapper[4696]: I0321 08:51:09.181264 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 21 08:51:09 crc kubenswrapper[4696]: I0321 08:51:09.488979 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-br9c4" event={"ID":"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a","Type":"ContainerStarted","Data":"3217cfe47d98f53b09d21bbc10fad865144d753bdccf955cd24aef22067aab64"} Mar 21 08:51:09 crc kubenswrapper[4696]: I0321 08:51:09.504685 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:51:09 crc kubenswrapper[4696]: I0321 08:51:09.505081 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" containerName="ceilometer-central-agent" containerID="cri-o://9d00b013cc0ef90677d79280d7ad247cccf2daf66eaf6f11cb75fd883f410c37" gracePeriod=30 Mar 21 08:51:09 crc kubenswrapper[4696]: I0321 08:51:09.505145 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" containerName="proxy-httpd" containerID="cri-o://f84e937cd55f41687d35f38f8a0e1f5b0d2a53c681b57c74d51a161e910e9ed5" gracePeriod=30 Mar 21 08:51:09 crc kubenswrapper[4696]: I0321 08:51:09.505204 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" containerName="sg-core" containerID="cri-o://c3316c4e2ea7b7880570f92fd7258613ef9e26551661cba16cd26594b8e33910" gracePeriod=30 Mar 21 08:51:09 crc kubenswrapper[4696]: I0321 08:51:09.505252 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" containerName="ceilometer-notification-agent" containerID="cri-o://04bd95501bb3b6bac5fa56ada1fc312ed788d6a33ec38d1182f877c9bc2eeed2" gracePeriod=30 Mar 21 08:51:10 crc kubenswrapper[4696]: I0321 08:51:10.513690 4696 generic.go:334] "Generic (PLEG): container finished" podID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" containerID="f84e937cd55f41687d35f38f8a0e1f5b0d2a53c681b57c74d51a161e910e9ed5" exitCode=0 Mar 21 08:51:10 crc kubenswrapper[4696]: I0321 08:51:10.513973 4696 generic.go:334] "Generic (PLEG): container finished" podID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" containerID="c3316c4e2ea7b7880570f92fd7258613ef9e26551661cba16cd26594b8e33910" exitCode=2 Mar 21 08:51:10 crc kubenswrapper[4696]: I0321 08:51:10.513983 4696 generic.go:334] "Generic (PLEG): container finished" podID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" containerID="9d00b013cc0ef90677d79280d7ad247cccf2daf66eaf6f11cb75fd883f410c37" exitCode=0 Mar 21 08:51:10 crc kubenswrapper[4696]: I0321 08:51:10.513858 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e6b9c7f-6556-4ab0-ac07-4c11a662000a","Type":"ContainerDied","Data":"f84e937cd55f41687d35f38f8a0e1f5b0d2a53c681b57c74d51a161e910e9ed5"} Mar 21 08:51:10 crc kubenswrapper[4696]: I0321 08:51:10.514568 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e6b9c7f-6556-4ab0-ac07-4c11a662000a","Type":"ContainerDied","Data":"c3316c4e2ea7b7880570f92fd7258613ef9e26551661cba16cd26594b8e33910"} Mar 21 08:51:10 crc kubenswrapper[4696]: I0321 08:51:10.514586 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e6b9c7f-6556-4ab0-ac07-4c11a662000a","Type":"ContainerDied","Data":"9d00b013cc0ef90677d79280d7ad247cccf2daf66eaf6f11cb75fd883f410c37"} Mar 21 08:51:13 crc kubenswrapper[4696]: I0321 08:51:13.127887 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" containerName="rabbitmq" containerID="cri-o://8be3270ef4edb543fbfd2456b063317996cad6cb123ab298417ce81d8c7a08c6" gracePeriod=604796 Mar 21 08:51:13 crc kubenswrapper[4696]: I0321 08:51:13.570321 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="b057ad21-8030-49e9-b400-c36d433b9f8c" containerName="rabbitmq" containerID="cri-o://b9bb5d044956676c0e0405393dd7743a5061cd8b121716b8f1aa7cd1c1987a17" gracePeriod=604796 Mar 21 08:51:16 crc kubenswrapper[4696]: I0321 08:51:16.087864 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="b057ad21-8030-49e9-b400-c36d433b9f8c" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.109:5671: connect: connection refused" Mar 21 08:51:16 crc kubenswrapper[4696]: I0321 08:51:16.387242 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.110:5671: connect: connection refused" Mar 21 08:51:16 crc kubenswrapper[4696]: I0321 08:51:16.593665 4696 generic.go:334] "Generic (PLEG): container finished" podID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" containerID="04bd95501bb3b6bac5fa56ada1fc312ed788d6a33ec38d1182f877c9bc2eeed2" exitCode=0 Mar 21 08:51:16 crc kubenswrapper[4696]: I0321 08:51:16.593986 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e6b9c7f-6556-4ab0-ac07-4c11a662000a","Type":"ContainerDied","Data":"04bd95501bb3b6bac5fa56ada1fc312ed788d6a33ec38d1182f877c9bc2eeed2"} Mar 21 08:51:18 crc kubenswrapper[4696]: I0321 08:51:18.618878 4696 generic.go:334] "Generic (PLEG): container finished" podID="817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a" containerID="3217cfe47d98f53b09d21bbc10fad865144d753bdccf955cd24aef22067aab64" exitCode=0 Mar 21 08:51:18 crc kubenswrapper[4696]: I0321 08:51:18.618928 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-br9c4" event={"ID":"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a","Type":"ContainerDied","Data":"3217cfe47d98f53b09d21bbc10fad865144d753bdccf955cd24aef22067aab64"} Mar 21 08:51:19 crc kubenswrapper[4696]: I0321 08:51:19.635173 4696 generic.go:334] "Generic (PLEG): container finished" podID="0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" containerID="8be3270ef4edb543fbfd2456b063317996cad6cb123ab298417ce81d8c7a08c6" exitCode=0 Mar 21 08:51:19 crc kubenswrapper[4696]: I0321 08:51:19.635246 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a","Type":"ContainerDied","Data":"8be3270ef4edb543fbfd2456b063317996cad6cb123ab298417ce81d8c7a08c6"} Mar 21 08:51:20 crc kubenswrapper[4696]: I0321 08:51:20.671369 4696 generic.go:334] "Generic (PLEG): container finished" podID="b057ad21-8030-49e9-b400-c36d433b9f8c" containerID="b9bb5d044956676c0e0405393dd7743a5061cd8b121716b8f1aa7cd1c1987a17" exitCode=0 Mar 21 08:51:20 crc kubenswrapper[4696]: I0321 08:51:20.671483 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b057ad21-8030-49e9-b400-c36d433b9f8c","Type":"ContainerDied","Data":"b9bb5d044956676c0e0405393dd7743a5061cd8b121716b8f1aa7cd1c1987a17"} Mar 21 08:51:24 crc kubenswrapper[4696]: I0321 08:51:24.892743 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.049179 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-scripts\") pod \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.049240 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-config-data\") pod \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.049327 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-combined-ca-bundle\") pod \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.049359 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-run-httpd\") pod \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.049434 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-log-httpd\") pod \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.049472 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-ceilometer-tls-certs\") pod \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.049493 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhj99\" (UniqueName: \"kubernetes.io/projected/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-kube-api-access-hhj99\") pod \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.049584 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-sg-core-conf-yaml\") pod \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\" (UID: \"1e6b9c7f-6556-4ab0-ac07-4c11a662000a\") " Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.050032 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1e6b9c7f-6556-4ab0-ac07-4c11a662000a" (UID: "1e6b9c7f-6556-4ab0-ac07-4c11a662000a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.050522 4696 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.051901 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1e6b9c7f-6556-4ab0-ac07-4c11a662000a" (UID: "1e6b9c7f-6556-4ab0-ac07-4c11a662000a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.061980 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-kube-api-access-hhj99" (OuterVolumeSpecName: "kube-api-access-hhj99") pod "1e6b9c7f-6556-4ab0-ac07-4c11a662000a" (UID: "1e6b9c7f-6556-4ab0-ac07-4c11a662000a"). InnerVolumeSpecName "kube-api-access-hhj99". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.078328 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-scripts" (OuterVolumeSpecName: "scripts") pod "1e6b9c7f-6556-4ab0-ac07-4c11a662000a" (UID: "1e6b9c7f-6556-4ab0-ac07-4c11a662000a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.109911 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1e6b9c7f-6556-4ab0-ac07-4c11a662000a" (UID: "1e6b9c7f-6556-4ab0-ac07-4c11a662000a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.151901 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.151930 4696 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.151940 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhj99\" (UniqueName: \"kubernetes.io/projected/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-kube-api-access-hhj99\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.151950 4696 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.156862 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "1e6b9c7f-6556-4ab0-ac07-4c11a662000a" (UID: "1e6b9c7f-6556-4ab0-ac07-4c11a662000a"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.185727 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e6b9c7f-6556-4ab0-ac07-4c11a662000a" (UID: "1e6b9c7f-6556-4ab0-ac07-4c11a662000a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.222022 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-config-data" (OuterVolumeSpecName: "config-data") pod "1e6b9c7f-6556-4ab0-ac07-4c11a662000a" (UID: "1e6b9c7f-6556-4ab0-ac07-4c11a662000a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.256187 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.256221 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.256235 4696 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e6b9c7f-6556-4ab0-ac07-4c11a662000a-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.727332 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e6b9c7f-6556-4ab0-ac07-4c11a662000a","Type":"ContainerDied","Data":"f53c5d75d83f40a10774ac7a9c7a0386687b3d7627fa966f16eb8892db66ff75"} Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.727393 4696 scope.go:117] "RemoveContainer" containerID="f84e937cd55f41687d35f38f8a0e1f5b0d2a53c681b57c74d51a161e910e9ed5" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.727419 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.786230 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.807872 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.824945 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:51:25 crc kubenswrapper[4696]: E0321 08:51:25.825526 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" containerName="sg-core" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.825548 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" containerName="sg-core" Mar 21 08:51:25 crc kubenswrapper[4696]: E0321 08:51:25.825566 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" containerName="proxy-httpd" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.825574 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" containerName="proxy-httpd" Mar 21 08:51:25 crc kubenswrapper[4696]: E0321 08:51:25.825609 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" containerName="ceilometer-central-agent" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.825617 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" containerName="ceilometer-central-agent" Mar 21 08:51:25 crc kubenswrapper[4696]: E0321 08:51:25.825630 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" containerName="ceilometer-notification-agent" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.825638 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" containerName="ceilometer-notification-agent" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.825891 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" containerName="ceilometer-notification-agent" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.825916 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" containerName="proxy-httpd" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.825945 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" containerName="sg-core" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.825955 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" containerName="ceilometer-central-agent" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.828378 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.830244 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.831434 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.831674 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.833085 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.977234 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-scripts\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.977623 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-log-httpd\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.977673 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wllfb\" (UniqueName: \"kubernetes.io/projected/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-kube-api-access-wllfb\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.977714 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.977803 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-run-httpd\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.977903 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.977977 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:25 crc kubenswrapper[4696]: I0321 08:51:25.978004 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-config-data\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.031365 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-dbb88bf8c-gjwx5"] Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.033893 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.038551 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.047101 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dbb88bf8c-gjwx5"] Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.079961 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wllfb\" (UniqueName: \"kubernetes.io/projected/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-kube-api-access-wllfb\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.080023 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.080099 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-run-httpd\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.080156 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.080203 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.080226 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-config-data\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.080284 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-scripts\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.080303 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-log-httpd\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.080762 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-log-httpd\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.084514 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-run-httpd\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.090658 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-scripts\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.092321 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.093494 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.101054 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-config-data\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.105577 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.124737 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wllfb\" (UniqueName: \"kubernetes.io/projected/72bd0225-bcbe-44ec-be2a-4f38093ea9c9-kube-api-access-wllfb\") pod \"ceilometer-0\" (UID: \"72bd0225-bcbe-44ec-be2a-4f38093ea9c9\") " pod="openstack/ceilometer-0" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.178493 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.183085 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-config\") pod \"dnsmasq-dns-dbb88bf8c-gjwx5\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.183136 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-dns-swift-storage-0\") pod \"dnsmasq-dns-dbb88bf8c-gjwx5\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.183159 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-dns-svc\") pod \"dnsmasq-dns-dbb88bf8c-gjwx5\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.183191 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-ovsdbserver-sb\") pod \"dnsmasq-dns-dbb88bf8c-gjwx5\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.183232 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-ovsdbserver-nb\") pod \"dnsmasq-dns-dbb88bf8c-gjwx5\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.183256 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkhbf\" (UniqueName: \"kubernetes.io/projected/1e77a97c-92ea-4bc5-bb15-b01d360220b6-kube-api-access-lkhbf\") pod \"dnsmasq-dns-dbb88bf8c-gjwx5\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.183308 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-openstack-edpm-ipam\") pod \"dnsmasq-dns-dbb88bf8c-gjwx5\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.285388 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-dns-swift-storage-0\") pod \"dnsmasq-dns-dbb88bf8c-gjwx5\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.285442 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-dns-svc\") pod \"dnsmasq-dns-dbb88bf8c-gjwx5\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.285487 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-ovsdbserver-sb\") pod \"dnsmasq-dns-dbb88bf8c-gjwx5\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.285553 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-ovsdbserver-nb\") pod \"dnsmasq-dns-dbb88bf8c-gjwx5\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.285587 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkhbf\" (UniqueName: \"kubernetes.io/projected/1e77a97c-92ea-4bc5-bb15-b01d360220b6-kube-api-access-lkhbf\") pod \"dnsmasq-dns-dbb88bf8c-gjwx5\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.285657 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-openstack-edpm-ipam\") pod \"dnsmasq-dns-dbb88bf8c-gjwx5\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.285771 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-config\") pod \"dnsmasq-dns-dbb88bf8c-gjwx5\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.286552 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-dns-svc\") pod \"dnsmasq-dns-dbb88bf8c-gjwx5\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.287029 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-config\") pod \"dnsmasq-dns-dbb88bf8c-gjwx5\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.287382 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-dns-swift-storage-0\") pod \"dnsmasq-dns-dbb88bf8c-gjwx5\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.287477 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-ovsdbserver-sb\") pod \"dnsmasq-dns-dbb88bf8c-gjwx5\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.287599 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-openstack-edpm-ipam\") pod \"dnsmasq-dns-dbb88bf8c-gjwx5\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.288023 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-ovsdbserver-nb\") pod \"dnsmasq-dns-dbb88bf8c-gjwx5\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.337267 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkhbf\" (UniqueName: \"kubernetes.io/projected/1e77a97c-92ea-4bc5-bb15-b01d360220b6-kube-api-access-lkhbf\") pod \"dnsmasq-dns-dbb88bf8c-gjwx5\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.358434 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:26 crc kubenswrapper[4696]: I0321 08:51:26.559208 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" path="/var/lib/kubelet/pods/1e6b9c7f-6556-4ab0-ac07-4c11a662000a/volumes" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.549402 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.549406 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.677948 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d\") pod \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.678009 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b057ad21-8030-49e9-b400-c36d433b9f8c-pod-info\") pod \"b057ad21-8030-49e9-b400-c36d433b9f8c\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.678078 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b057ad21-8030-49e9-b400-c36d433b9f8c-plugins-conf\") pod \"b057ad21-8030-49e9-b400-c36d433b9f8c\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.681726 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b057ad21-8030-49e9-b400-c36d433b9f8c-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "b057ad21-8030-49e9-b400-c36d433b9f8c" (UID: "b057ad21-8030-49e9-b400-c36d433b9f8c"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.686869 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f47b560c-4e99-44f7-928a-39b1f9203a25\") pod \"b057ad21-8030-49e9-b400-c36d433b9f8c\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.686963 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-confd\") pod \"b057ad21-8030-49e9-b400-c36d433b9f8c\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.686997 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-confd\") pod \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.687104 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-server-conf\") pod \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.687156 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-config-data\") pod \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.687188 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b057ad21-8030-49e9-b400-c36d433b9f8c-erlang-cookie-secret\") pod \"b057ad21-8030-49e9-b400-c36d433b9f8c\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.687251 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-tls\") pod \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.687289 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-pod-info\") pod \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.687319 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-erlang-cookie\") pod \"b057ad21-8030-49e9-b400-c36d433b9f8c\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.687344 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-plugins\") pod \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.687371 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b057ad21-8030-49e9-b400-c36d433b9f8c-config-data\") pod \"b057ad21-8030-49e9-b400-c36d433b9f8c\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.687406 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-tls\") pod \"b057ad21-8030-49e9-b400-c36d433b9f8c\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.687428 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-erlang-cookie-secret\") pod \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.687471 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjtkr\" (UniqueName: \"kubernetes.io/projected/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-kube-api-access-gjtkr\") pod \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.687511 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-plugins\") pod \"b057ad21-8030-49e9-b400-c36d433b9f8c\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.687545 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-erlang-cookie\") pod \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.687585 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b057ad21-8030-49e9-b400-c36d433b9f8c-server-conf\") pod \"b057ad21-8030-49e9-b400-c36d433b9f8c\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.694564 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-plugins-conf\") pod \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\" (UID: \"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.694618 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmw6t\" (UniqueName: \"kubernetes.io/projected/b057ad21-8030-49e9-b400-c36d433b9f8c-kube-api-access-dmw6t\") pod \"b057ad21-8030-49e9-b400-c36d433b9f8c\" (UID: \"b057ad21-8030-49e9-b400-c36d433b9f8c\") " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.690269 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" (UID: "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.692779 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/b057ad21-8030-49e9-b400-c36d433b9f8c-pod-info" (OuterVolumeSpecName: "pod-info") pod "b057ad21-8030-49e9-b400-c36d433b9f8c" (UID: "b057ad21-8030-49e9-b400-c36d433b9f8c"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.693415 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "b057ad21-8030-49e9-b400-c36d433b9f8c" (UID: "b057ad21-8030-49e9-b400-c36d433b9f8c"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.695498 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "b057ad21-8030-49e9-b400-c36d433b9f8c" (UID: "b057ad21-8030-49e9-b400-c36d433b9f8c"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.695880 4696 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b057ad21-8030-49e9-b400-c36d433b9f8c-pod-info\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.695898 4696 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b057ad21-8030-49e9-b400-c36d433b9f8c-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.695908 4696 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.695917 4696 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.695925 4696 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.696899 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" (UID: "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.700168 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" (UID: "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.718210 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "b057ad21-8030-49e9-b400-c36d433b9f8c" (UID: "b057ad21-8030-49e9-b400-c36d433b9f8c"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.719362 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b057ad21-8030-49e9-b400-c36d433b9f8c-kube-api-access-dmw6t" (OuterVolumeSpecName: "kube-api-access-dmw6t") pod "b057ad21-8030-49e9-b400-c36d433b9f8c" (UID: "b057ad21-8030-49e9-b400-c36d433b9f8c"). InnerVolumeSpecName "kube-api-access-dmw6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.719462 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-pod-info" (OuterVolumeSpecName: "pod-info") pod "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" (UID: "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.722978 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" (UID: "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.748091 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" (UID: "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.748188 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b057ad21-8030-49e9-b400-c36d433b9f8c-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "b057ad21-8030-49e9-b400-c36d433b9f8c" (UID: "b057ad21-8030-49e9-b400-c36d433b9f8c"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.751244 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-kube-api-access-gjtkr" (OuterVolumeSpecName: "kube-api-access-gjtkr") pod "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" (UID: "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a"). InnerVolumeSpecName "kube-api-access-gjtkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.792652 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b057ad21-8030-49e9-b400-c36d433b9f8c-config-data" (OuterVolumeSpecName: "config-data") pod "b057ad21-8030-49e9-b400-c36d433b9f8c" (UID: "b057ad21-8030-49e9-b400-c36d433b9f8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.802964 4696 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.803197 4696 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-pod-info\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.803276 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b057ad21-8030-49e9-b400-c36d433b9f8c-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.803358 4696 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.803435 4696 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.808927 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjtkr\" (UniqueName: \"kubernetes.io/projected/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-kube-api-access-gjtkr\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.809546 4696 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.809694 4696 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.809791 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmw6t\" (UniqueName: \"kubernetes.io/projected/b057ad21-8030-49e9-b400-c36d433b9f8c-kube-api-access-dmw6t\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.809910 4696 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b057ad21-8030-49e9-b400-c36d433b9f8c-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.849287 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d" (OuterVolumeSpecName: "persistence") pod "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" (UID: "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a"). InnerVolumeSpecName "pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.849300 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-config-data" (OuterVolumeSpecName: "config-data") pod "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" (UID: "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.854962 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b057ad21-8030-49e9-b400-c36d433b9f8c","Type":"ContainerDied","Data":"8c465eaf26c5aa0886cd99d3f25096099b9e93a7f62b3cae2b59029502c9e221"} Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.855143 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.862022 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f47b560c-4e99-44f7-928a-39b1f9203a25" (OuterVolumeSpecName: "persistence") pod "b057ad21-8030-49e9-b400-c36d433b9f8c" (UID: "b057ad21-8030-49e9-b400-c36d433b9f8c"). InnerVolumeSpecName "pvc-f47b560c-4e99-44f7-928a-39b1f9203a25". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.864954 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a","Type":"ContainerDied","Data":"589d42286ee21cc59f6a9512729069f2eff8a5e35c1b64b7c8364d3eb6a79e09"} Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.865764 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.871507 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b057ad21-8030-49e9-b400-c36d433b9f8c-server-conf" (OuterVolumeSpecName: "server-conf") pod "b057ad21-8030-49e9-b400-c36d433b9f8c" (UID: "b057ad21-8030-49e9-b400-c36d433b9f8c"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.876671 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-server-conf" (OuterVolumeSpecName: "server-conf") pod "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" (UID: "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.911852 4696 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b057ad21-8030-49e9-b400-c36d433b9f8c-server-conf\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.911901 4696 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d\") on node \"crc\" " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.911922 4696 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-f47b560c-4e99-44f7-928a-39b1f9203a25\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f47b560c-4e99-44f7-928a-39b1f9203a25\") on node \"crc\" " Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.911937 4696 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-server-conf\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.911949 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.943734 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "b057ad21-8030-49e9-b400-c36d433b9f8c" (UID: "b057ad21-8030-49e9-b400-c36d433b9f8c"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.955109 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" (UID: "0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.975626 4696 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.975971 4696 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-f47b560c-4e99-44f7-928a-39b1f9203a25" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f47b560c-4e99-44f7-928a-39b1f9203a25") on node "crc" Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.976201 4696 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 21 08:51:29 crc kubenswrapper[4696]: I0321 08:51:29.977112 4696 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d") on node "crc" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.014971 4696 reconciler_common.go:293] "Volume detached for volume \"pvc-f47b560c-4e99-44f7-928a-39b1f9203a25\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f47b560c-4e99-44f7-928a-39b1f9203a25\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.015039 4696 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b057ad21-8030-49e9-b400-c36d433b9f8c-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.015056 4696 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.015071 4696 reconciler_common.go:293] "Volume detached for volume \"pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.207504 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.228035 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.246457 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.263615 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.278102 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 21 08:51:30 crc kubenswrapper[4696]: E0321 08:51:30.278566 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" containerName="rabbitmq" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.278583 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" containerName="rabbitmq" Mar 21 08:51:30 crc kubenswrapper[4696]: E0321 08:51:30.278605 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b057ad21-8030-49e9-b400-c36d433b9f8c" containerName="rabbitmq" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.278612 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b057ad21-8030-49e9-b400-c36d433b9f8c" containerName="rabbitmq" Mar 21 08:51:30 crc kubenswrapper[4696]: E0321 08:51:30.278623 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b057ad21-8030-49e9-b400-c36d433b9f8c" containerName="setup-container" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.278632 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b057ad21-8030-49e9-b400-c36d433b9f8c" containerName="setup-container" Mar 21 08:51:30 crc kubenswrapper[4696]: E0321 08:51:30.278876 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" containerName="setup-container" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.278885 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" containerName="setup-container" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.279094 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" containerName="rabbitmq" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.279114 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="b057ad21-8030-49e9-b400-c36d433b9f8c" containerName="rabbitmq" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.280437 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.283829 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.284080 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-mxjmk" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.284186 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.284188 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.284295 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.284382 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.284436 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.290260 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.292938 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.301498 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.301502 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.302115 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.302432 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.302805 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.303056 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-hzkmd" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.303331 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.310381 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.320634 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.323349 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0f40066c-3b43-42a5-9bb2-c8fb94a921db-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.323418 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f47b560c-4e99-44f7-928a-39b1f9203a25\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f47b560c-4e99-44f7-928a-39b1f9203a25\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.323523 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4cjk\" (UniqueName: \"kubernetes.io/projected/0f40066c-3b43-42a5-9bb2-c8fb94a921db-kube-api-access-f4cjk\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.323562 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0f40066c-3b43-42a5-9bb2-c8fb94a921db-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.323746 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0f40066c-3b43-42a5-9bb2-c8fb94a921db-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.323784 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0f40066c-3b43-42a5-9bb2-c8fb94a921db-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.323890 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0f40066c-3b43-42a5-9bb2-c8fb94a921db-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.323925 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0f40066c-3b43-42a5-9bb2-c8fb94a921db-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.323953 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0f40066c-3b43-42a5-9bb2-c8fb94a921db-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.323978 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f40066c-3b43-42a5-9bb2-c8fb94a921db-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.324047 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0f40066c-3b43-42a5-9bb2-c8fb94a921db-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.340931 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.340982 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.341024 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.341571 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"25b5c4ade4d0fb1258c68b7e200638ed962f1ba1ce964274a9e7589cd0a163c2"} pod="openshift-machine-config-operator/machine-config-daemon-z7srw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.341622 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" containerID="cri-o://25b5c4ade4d0fb1258c68b7e200638ed962f1ba1ce964274a9e7589cd0a163c2" gracePeriod=600 Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.427020 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/56d6eab2-0e95-402a-afb7-2707ca41b144-server-conf\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.427076 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/56d6eab2-0e95-402a-afb7-2707ca41b144-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.427127 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0f40066c-3b43-42a5-9bb2-c8fb94a921db-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.427198 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0f40066c-3b43-42a5-9bb2-c8fb94a921db-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.427241 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0f40066c-3b43-42a5-9bb2-c8fb94a921db-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.427269 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f40066c-3b43-42a5-9bb2-c8fb94a921db-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.427333 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/56d6eab2-0e95-402a-afb7-2707ca41b144-pod-info\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.427376 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0f40066c-3b43-42a5-9bb2-c8fb94a921db-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.427414 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ts8sh\" (UniqueName: \"kubernetes.io/projected/56d6eab2-0e95-402a-afb7-2707ca41b144-kube-api-access-ts8sh\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.427470 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/56d6eab2-0e95-402a-afb7-2707ca41b144-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.427519 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0f40066c-3b43-42a5-9bb2-c8fb94a921db-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.427578 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.427614 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56d6eab2-0e95-402a-afb7-2707ca41b144-config-data\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.427648 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f47b560c-4e99-44f7-928a-39b1f9203a25\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f47b560c-4e99-44f7-928a-39b1f9203a25\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.427693 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/56d6eab2-0e95-402a-afb7-2707ca41b144-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.427716 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/56d6eab2-0e95-402a-afb7-2707ca41b144-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.427748 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4cjk\" (UniqueName: \"kubernetes.io/projected/0f40066c-3b43-42a5-9bb2-c8fb94a921db-kube-api-access-f4cjk\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.427799 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0f40066c-3b43-42a5-9bb2-c8fb94a921db-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.428072 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/56d6eab2-0e95-402a-afb7-2707ca41b144-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.428151 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0f40066c-3b43-42a5-9bb2-c8fb94a921db-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.428177 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/56d6eab2-0e95-402a-afb7-2707ca41b144-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.428212 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0f40066c-3b43-42a5-9bb2-c8fb94a921db-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.430187 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0f40066c-3b43-42a5-9bb2-c8fb94a921db-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.430197 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0f40066c-3b43-42a5-9bb2-c8fb94a921db-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.430798 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f40066c-3b43-42a5-9bb2-c8fb94a921db-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.431447 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0f40066c-3b43-42a5-9bb2-c8fb94a921db-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.431502 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0f40066c-3b43-42a5-9bb2-c8fb94a921db-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.431947 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0f40066c-3b43-42a5-9bb2-c8fb94a921db-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.435672 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0f40066c-3b43-42a5-9bb2-c8fb94a921db-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.454104 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0f40066c-3b43-42a5-9bb2-c8fb94a921db-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.459858 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.459903 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f47b560c-4e99-44f7-928a-39b1f9203a25\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f47b560c-4e99-44f7-928a-39b1f9203a25\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/3d548c5607941a22cfe7a5753cb098dc630cefe69835678ddcd11d0cea80b16b/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.460638 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4cjk\" (UniqueName: \"kubernetes.io/projected/0f40066c-3b43-42a5-9bb2-c8fb94a921db-kube-api-access-f4cjk\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.461140 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0f40066c-3b43-42a5-9bb2-c8fb94a921db-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.530025 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/56d6eab2-0e95-402a-afb7-2707ca41b144-pod-info\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.530097 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ts8sh\" (UniqueName: \"kubernetes.io/projected/56d6eab2-0e95-402a-afb7-2707ca41b144-kube-api-access-ts8sh\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.530142 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/56d6eab2-0e95-402a-afb7-2707ca41b144-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.530215 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.530249 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56d6eab2-0e95-402a-afb7-2707ca41b144-config-data\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.530287 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/56d6eab2-0e95-402a-afb7-2707ca41b144-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.530308 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/56d6eab2-0e95-402a-afb7-2707ca41b144-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.530402 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/56d6eab2-0e95-402a-afb7-2707ca41b144-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.530450 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/56d6eab2-0e95-402a-afb7-2707ca41b144-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.530505 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/56d6eab2-0e95-402a-afb7-2707ca41b144-server-conf\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.530533 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/56d6eab2-0e95-402a-afb7-2707ca41b144-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.531445 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/56d6eab2-0e95-402a-afb7-2707ca41b144-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.534215 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/56d6eab2-0e95-402a-afb7-2707ca41b144-server-conf\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.534676 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/56d6eab2-0e95-402a-afb7-2707ca41b144-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.534902 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/56d6eab2-0e95-402a-afb7-2707ca41b144-pod-info\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.535291 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56d6eab2-0e95-402a-afb7-2707ca41b144-config-data\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.535313 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/56d6eab2-0e95-402a-afb7-2707ca41b144-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.535650 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/56d6eab2-0e95-402a-afb7-2707ca41b144-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.537001 4696 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.537037 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/087398aff111d894aab0c73f079ac83ad54622acef4ed647806a18bd4dd4491c/globalmount\"" pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.539096 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/56d6eab2-0e95-402a-afb7-2707ca41b144-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.543079 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/56d6eab2-0e95-402a-afb7-2707ca41b144-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.555015 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f47b560c-4e99-44f7-928a-39b1f9203a25\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f47b560c-4e99-44f7-928a-39b1f9203a25\") pod \"rabbitmq-cell1-server-0\" (UID: \"0f40066c-3b43-42a5-9bb2-c8fb94a921db\") " pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.558576 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ts8sh\" (UniqueName: \"kubernetes.io/projected/56d6eab2-0e95-402a-afb7-2707ca41b144-kube-api-access-ts8sh\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.581719 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" path="/var/lib/kubelet/pods/0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a/volumes" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.582674 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b057ad21-8030-49e9-b400-c36d433b9f8c" path="/var/lib/kubelet/pods/b057ad21-8030-49e9-b400-c36d433b9f8c/volumes" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.618301 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.620447 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2907fb4b-d32d-4f82-839b-fb05b66c7e6d\") pod \"rabbitmq-server-0\" (UID: \"56d6eab2-0e95-402a-afb7-2707ca41b144\") " pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.639643 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.886145 4696 generic.go:334] "Generic (PLEG): container finished" podID="daaf227a-2305-495e-8495-a280abcd8e10" containerID="25b5c4ade4d0fb1258c68b7e200638ed962f1ba1ce964274a9e7589cd0a163c2" exitCode=0 Mar 21 08:51:30 crc kubenswrapper[4696]: I0321 08:51:30.886196 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerDied","Data":"25b5c4ade4d0fb1258c68b7e200638ed962f1ba1ce964274a9e7589cd0a163c2"} Mar 21 08:51:31 crc kubenswrapper[4696]: I0321 08:51:31.087569 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="b057ad21-8030-49e9-b400-c36d433b9f8c" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.109:5671: i/o timeout" Mar 21 08:51:31 crc kubenswrapper[4696]: I0321 08:51:31.153357 4696 scope.go:117] "RemoveContainer" containerID="c3316c4e2ea7b7880570f92fd7258613ef9e26551661cba16cd26594b8e33910" Mar 21 08:51:31 crc kubenswrapper[4696]: E0321 08:51:31.194378 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current-tested" Mar 21 08:51:31 crc kubenswrapper[4696]: E0321 08:51:31.194417 4696 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current-tested" Mar 21 08:51:31 crc kubenswrapper[4696]: E0321 08:51:31.194538 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cloudkitty-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CloudKittyPassword,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:CloudKittyPassword,Optional:nil,},},},EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:cloudkitty-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:certs,ReadOnly:true,MountPath:/var/lib/openstack/loki-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5kzqv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42406,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-db-sync-fq9xb_openstack(1f8a2d50-6415-4d3a-8cb4-bbbfce785842): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 08:51:31 crc kubenswrapper[4696]: E0321 08:51:31.196630 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cloudkitty-db-sync-fq9xb" podUID="1f8a2d50-6415-4d3a-8cb4-bbbfce785842" Mar 21 08:51:31 crc kubenswrapper[4696]: I0321 08:51:31.288177 4696 scope.go:117] "RemoveContainer" containerID="04bd95501bb3b6bac5fa56ada1fc312ed788d6a33ec38d1182f877c9bc2eeed2" Mar 21 08:51:31 crc kubenswrapper[4696]: I0321 08:51:31.389720 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="0ea6ac8c-cb0e-46a5-b3a2-0701bd7dce0a" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.110:5671: i/o timeout" Mar 21 08:51:31 crc kubenswrapper[4696]: I0321 08:51:31.437461 4696 scope.go:117] "RemoveContainer" containerID="9d00b013cc0ef90677d79280d7ad247cccf2daf66eaf6f11cb75fd883f410c37" Mar 21 08:51:31 crc kubenswrapper[4696]: I0321 08:51:31.687990 4696 scope.go:117] "RemoveContainer" containerID="b9bb5d044956676c0e0405393dd7743a5061cd8b121716b8f1aa7cd1c1987a17" Mar 21 08:51:31 crc kubenswrapper[4696]: I0321 08:51:31.733030 4696 scope.go:117] "RemoveContainer" containerID="7fd27604a7d301ac11fa0074a203113325187e8f947c4308c11343eb67ceaed5" Mar 21 08:51:31 crc kubenswrapper[4696]: I0321 08:51:31.772221 4696 scope.go:117] "RemoveContainer" containerID="8be3270ef4edb543fbfd2456b063317996cad6cb123ab298417ce81d8c7a08c6" Mar 21 08:51:31 crc kubenswrapper[4696]: I0321 08:51:31.825390 4696 scope.go:117] "RemoveContainer" containerID="e8e800b9e6d78c36e85ad2b994e32f91f76eb2ffdc75250bf0f1b501cb1813a8" Mar 21 08:51:31 crc kubenswrapper[4696]: I0321 08:51:31.870411 4696 scope.go:117] "RemoveContainer" containerID="37dc28f551f9db6ffb052a7a69754f617535fa9f02cee7492942ac3ed542a742" Mar 21 08:51:31 crc kubenswrapper[4696]: I0321 08:51:31.910501 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerStarted","Data":"a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959"} Mar 21 08:51:31 crc kubenswrapper[4696]: I0321 08:51:31.915108 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-br9c4" event={"ID":"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a","Type":"ContainerStarted","Data":"f1f42ab54713d20c94a91d51039358cfd6b2751fbb78b7b016e31b7602c81ee0"} Mar 21 08:51:31 crc kubenswrapper[4696]: E0321 08:51:31.926915 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current-tested\\\"\"" pod="openstack/cloudkitty-db-sync-fq9xb" podUID="1f8a2d50-6415-4d3a-8cb4-bbbfce785842" Mar 21 08:51:31 crc kubenswrapper[4696]: I0321 08:51:31.996318 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-br9c4" podStartSLOduration=3.192447116 podStartE2EDuration="25.996296228s" podCreationTimestamp="2026-03-21 08:51:06 +0000 UTC" firstStartedPulling="2026-03-21 08:51:08.458197448 +0000 UTC m=+1402.579078161" lastFinishedPulling="2026-03-21 08:51:31.26204656 +0000 UTC m=+1425.382927273" observedRunningTime="2026-03-21 08:51:31.966150592 +0000 UTC m=+1426.087031305" watchObservedRunningTime="2026-03-21 08:51:31.996296228 +0000 UTC m=+1426.117176941" Mar 21 08:51:32 crc kubenswrapper[4696]: I0321 08:51:32.304604 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dbb88bf8c-gjwx5"] Mar 21 08:51:32 crc kubenswrapper[4696]: I0321 08:51:32.354119 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 21 08:51:32 crc kubenswrapper[4696]: I0321 08:51:32.449666 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 21 08:51:32 crc kubenswrapper[4696]: I0321 08:51:32.657202 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 21 08:51:32 crc kubenswrapper[4696]: I0321 08:51:32.955263 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72bd0225-bcbe-44ec-be2a-4f38093ea9c9","Type":"ContainerStarted","Data":"85ed305a44859ea16c4c3051c9692139db2ffd8094df1f8350d5d67294e9766a"} Mar 21 08:51:32 crc kubenswrapper[4696]: I0321 08:51:32.958411 4696 generic.go:334] "Generic (PLEG): container finished" podID="1e77a97c-92ea-4bc5-bb15-b01d360220b6" containerID="56e2fc9386018f8d98cb4e8c2eedfd1c0d6cd7f88c2355450d0bc4edac6c3379" exitCode=0 Mar 21 08:51:32 crc kubenswrapper[4696]: I0321 08:51:32.958481 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" event={"ID":"1e77a97c-92ea-4bc5-bb15-b01d360220b6","Type":"ContainerDied","Data":"56e2fc9386018f8d98cb4e8c2eedfd1c0d6cd7f88c2355450d0bc4edac6c3379"} Mar 21 08:51:32 crc kubenswrapper[4696]: I0321 08:51:32.958526 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" event={"ID":"1e77a97c-92ea-4bc5-bb15-b01d360220b6","Type":"ContainerStarted","Data":"cebca7f8cba161f8e5e5e0bff62e1a7d4c22d953179115d57357f890b77aa577"} Mar 21 08:51:32 crc kubenswrapper[4696]: I0321 08:51:32.968529 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"56d6eab2-0e95-402a-afb7-2707ca41b144","Type":"ContainerStarted","Data":"482c440f185ec1f7c4044fa8621e6dda6350524e8898dacc12920857b3c28a1a"} Mar 21 08:51:32 crc kubenswrapper[4696]: I0321 08:51:32.973727 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0f40066c-3b43-42a5-9bb2-c8fb94a921db","Type":"ContainerStarted","Data":"7f884f3f99f6e45a5a242a089bdda2261f3284acc27d2c0b2aea76d9102b587e"} Mar 21 08:51:33 crc kubenswrapper[4696]: I0321 08:51:33.985714 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" event={"ID":"1e77a97c-92ea-4bc5-bb15-b01d360220b6","Type":"ContainerStarted","Data":"6901cb5ddec0972f962e20bf62f5fbd48b07096d1d54cf8ac35af0ead3781b8b"} Mar 21 08:51:33 crc kubenswrapper[4696]: I0321 08:51:33.987634 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:34 crc kubenswrapper[4696]: I0321 08:51:34.022317 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" podStartSLOduration=8.022300491 podStartE2EDuration="8.022300491s" podCreationTimestamp="2026-03-21 08:51:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:51:34.015024566 +0000 UTC m=+1428.135905279" watchObservedRunningTime="2026-03-21 08:51:34.022300491 +0000 UTC m=+1428.143181194" Mar 21 08:51:34 crc kubenswrapper[4696]: I0321 08:51:34.996301 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"56d6eab2-0e95-402a-afb7-2707ca41b144","Type":"ContainerStarted","Data":"80bc32955c6457e7b25214605028e3fda13557029d2ab6eb5265f8915dc62edc"} Mar 21 08:51:34 crc kubenswrapper[4696]: I0321 08:51:34.997503 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0f40066c-3b43-42a5-9bb2-c8fb94a921db","Type":"ContainerStarted","Data":"591cc92872a6b54cb31429b395e1235f349c706bd3614cafb673fdfb11a4b1f1"} Mar 21 08:51:36 crc kubenswrapper[4696]: I0321 08:51:36.956242 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-br9c4" Mar 21 08:51:36 crc kubenswrapper[4696]: I0321 08:51:36.956986 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-br9c4" Mar 21 08:51:38 crc kubenswrapper[4696]: I0321 08:51:38.013003 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-br9c4" podUID="817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a" containerName="registry-server" probeResult="failure" output=< Mar 21 08:51:38 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Mar 21 08:51:38 crc kubenswrapper[4696]: > Mar 21 08:51:38 crc kubenswrapper[4696]: I0321 08:51:38.027354 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72bd0225-bcbe-44ec-be2a-4f38093ea9c9","Type":"ContainerStarted","Data":"737dfe84b5fb043a47dc78381f9514256ce9cf3868475cbf9f13994ae1a8e0d9"} Mar 21 08:51:39 crc kubenswrapper[4696]: I0321 08:51:39.043183 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72bd0225-bcbe-44ec-be2a-4f38093ea9c9","Type":"ContainerStarted","Data":"1a2d8b90f79885cc79999fb9a8facaa1c24380b17db5727e4f1620e9a16fd6b3"} Mar 21 08:51:40 crc kubenswrapper[4696]: I0321 08:51:40.055666 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72bd0225-bcbe-44ec-be2a-4f38093ea9c9","Type":"ContainerStarted","Data":"45990d80e3971b5a8ce5c5ec3c04edfd01051337fcacf620c8f582eb969dfae2"} Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.069281 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72bd0225-bcbe-44ec-be2a-4f38093ea9c9","Type":"ContainerStarted","Data":"23184d49139cb5312ca0a590bd82d5c6f163afa3c3218ebbd5b4da20de5065af"} Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.070845 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.359985 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.384867 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=8.01130002 podStartE2EDuration="16.38484563s" podCreationTimestamp="2026-03-21 08:51:25 +0000 UTC" firstStartedPulling="2026-03-21 08:51:32.328339002 +0000 UTC m=+1426.449219715" lastFinishedPulling="2026-03-21 08:51:40.701884612 +0000 UTC m=+1434.822765325" observedRunningTime="2026-03-21 08:51:41.110122895 +0000 UTC m=+1435.231003628" watchObservedRunningTime="2026-03-21 08:51:41.38484563 +0000 UTC m=+1435.505726343" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.432580 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5fd9b586ff-xr8tz"] Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.433062 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" podUID="4acec0f5-fbf7-4900-92aa-181b72767e79" containerName="dnsmasq-dns" containerID="cri-o://98c8fe9d6eb7efe4cbb38d9718d2575729319445cfe0682e2f960ac47a953850" gracePeriod=10 Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.607765 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85f64749dc-tqgff"] Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.630324 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.640162 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85f64749dc-tqgff"] Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.726096 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9fe9d98-b02d-4fdd-918e-c296d52f7efd-config\") pod \"dnsmasq-dns-85f64749dc-tqgff\" (UID: \"c9fe9d98-b02d-4fdd-918e-c296d52f7efd\") " pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.726168 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9fe9d98-b02d-4fdd-918e-c296d52f7efd-dns-svc\") pod \"dnsmasq-dns-85f64749dc-tqgff\" (UID: \"c9fe9d98-b02d-4fdd-918e-c296d52f7efd\") " pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.726267 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9fe9d98-b02d-4fdd-918e-c296d52f7efd-ovsdbserver-sb\") pod \"dnsmasq-dns-85f64749dc-tqgff\" (UID: \"c9fe9d98-b02d-4fdd-918e-c296d52f7efd\") " pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.726322 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c9fe9d98-b02d-4fdd-918e-c296d52f7efd-openstack-edpm-ipam\") pod \"dnsmasq-dns-85f64749dc-tqgff\" (UID: \"c9fe9d98-b02d-4fdd-918e-c296d52f7efd\") " pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.726347 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9fe9d98-b02d-4fdd-918e-c296d52f7efd-dns-swift-storage-0\") pod \"dnsmasq-dns-85f64749dc-tqgff\" (UID: \"c9fe9d98-b02d-4fdd-918e-c296d52f7efd\") " pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.726369 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9fe9d98-b02d-4fdd-918e-c296d52f7efd-ovsdbserver-nb\") pod \"dnsmasq-dns-85f64749dc-tqgff\" (UID: \"c9fe9d98-b02d-4fdd-918e-c296d52f7efd\") " pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.726398 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddkm9\" (UniqueName: \"kubernetes.io/projected/c9fe9d98-b02d-4fdd-918e-c296d52f7efd-kube-api-access-ddkm9\") pod \"dnsmasq-dns-85f64749dc-tqgff\" (UID: \"c9fe9d98-b02d-4fdd-918e-c296d52f7efd\") " pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.830299 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9fe9d98-b02d-4fdd-918e-c296d52f7efd-ovsdbserver-nb\") pod \"dnsmasq-dns-85f64749dc-tqgff\" (UID: \"c9fe9d98-b02d-4fdd-918e-c296d52f7efd\") " pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.830602 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddkm9\" (UniqueName: \"kubernetes.io/projected/c9fe9d98-b02d-4fdd-918e-c296d52f7efd-kube-api-access-ddkm9\") pod \"dnsmasq-dns-85f64749dc-tqgff\" (UID: \"c9fe9d98-b02d-4fdd-918e-c296d52f7efd\") " pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.830672 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9fe9d98-b02d-4fdd-918e-c296d52f7efd-config\") pod \"dnsmasq-dns-85f64749dc-tqgff\" (UID: \"c9fe9d98-b02d-4fdd-918e-c296d52f7efd\") " pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.830716 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9fe9d98-b02d-4fdd-918e-c296d52f7efd-dns-svc\") pod \"dnsmasq-dns-85f64749dc-tqgff\" (UID: \"c9fe9d98-b02d-4fdd-918e-c296d52f7efd\") " pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.830798 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9fe9d98-b02d-4fdd-918e-c296d52f7efd-ovsdbserver-sb\") pod \"dnsmasq-dns-85f64749dc-tqgff\" (UID: \"c9fe9d98-b02d-4fdd-918e-c296d52f7efd\") " pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.830836 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c9fe9d98-b02d-4fdd-918e-c296d52f7efd-openstack-edpm-ipam\") pod \"dnsmasq-dns-85f64749dc-tqgff\" (UID: \"c9fe9d98-b02d-4fdd-918e-c296d52f7efd\") " pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.830859 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9fe9d98-b02d-4fdd-918e-c296d52f7efd-dns-swift-storage-0\") pod \"dnsmasq-dns-85f64749dc-tqgff\" (UID: \"c9fe9d98-b02d-4fdd-918e-c296d52f7efd\") " pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.831627 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9fe9d98-b02d-4fdd-918e-c296d52f7efd-config\") pod \"dnsmasq-dns-85f64749dc-tqgff\" (UID: \"c9fe9d98-b02d-4fdd-918e-c296d52f7efd\") " pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.832849 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9fe9d98-b02d-4fdd-918e-c296d52f7efd-dns-swift-storage-0\") pod \"dnsmasq-dns-85f64749dc-tqgff\" (UID: \"c9fe9d98-b02d-4fdd-918e-c296d52f7efd\") " pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.832999 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9fe9d98-b02d-4fdd-918e-c296d52f7efd-ovsdbserver-nb\") pod \"dnsmasq-dns-85f64749dc-tqgff\" (UID: \"c9fe9d98-b02d-4fdd-918e-c296d52f7efd\") " pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.837570 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c9fe9d98-b02d-4fdd-918e-c296d52f7efd-openstack-edpm-ipam\") pod \"dnsmasq-dns-85f64749dc-tqgff\" (UID: \"c9fe9d98-b02d-4fdd-918e-c296d52f7efd\") " pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.840070 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9fe9d98-b02d-4fdd-918e-c296d52f7efd-ovsdbserver-sb\") pod \"dnsmasq-dns-85f64749dc-tqgff\" (UID: \"c9fe9d98-b02d-4fdd-918e-c296d52f7efd\") " pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.840207 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9fe9d98-b02d-4fdd-918e-c296d52f7efd-dns-svc\") pod \"dnsmasq-dns-85f64749dc-tqgff\" (UID: \"c9fe9d98-b02d-4fdd-918e-c296d52f7efd\") " pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.863902 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddkm9\" (UniqueName: \"kubernetes.io/projected/c9fe9d98-b02d-4fdd-918e-c296d52f7efd-kube-api-access-ddkm9\") pod \"dnsmasq-dns-85f64749dc-tqgff\" (UID: \"c9fe9d98-b02d-4fdd-918e-c296d52f7efd\") " pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:41 crc kubenswrapper[4696]: I0321 08:51:41.969651 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.145024 4696 generic.go:334] "Generic (PLEG): container finished" podID="4acec0f5-fbf7-4900-92aa-181b72767e79" containerID="98c8fe9d6eb7efe4cbb38d9718d2575729319445cfe0682e2f960ac47a953850" exitCode=0 Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.146121 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" event={"ID":"4acec0f5-fbf7-4900-92aa-181b72767e79","Type":"ContainerDied","Data":"98c8fe9d6eb7efe4cbb38d9718d2575729319445cfe0682e2f960ac47a953850"} Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.600405 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.617291 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85f64749dc-tqgff"] Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.673507 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-dns-swift-storage-0\") pod \"4acec0f5-fbf7-4900-92aa-181b72767e79\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.674010 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-ovsdbserver-nb\") pod \"4acec0f5-fbf7-4900-92aa-181b72767e79\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.674145 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-dns-svc\") pod \"4acec0f5-fbf7-4900-92aa-181b72767e79\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.674188 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-ovsdbserver-sb\") pod \"4acec0f5-fbf7-4900-92aa-181b72767e79\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.674233 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-config\") pod \"4acec0f5-fbf7-4900-92aa-181b72767e79\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.674261 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sllqw\" (UniqueName: \"kubernetes.io/projected/4acec0f5-fbf7-4900-92aa-181b72767e79-kube-api-access-sllqw\") pod \"4acec0f5-fbf7-4900-92aa-181b72767e79\" (UID: \"4acec0f5-fbf7-4900-92aa-181b72767e79\") " Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.690661 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4acec0f5-fbf7-4900-92aa-181b72767e79-kube-api-access-sllqw" (OuterVolumeSpecName: "kube-api-access-sllqw") pod "4acec0f5-fbf7-4900-92aa-181b72767e79" (UID: "4acec0f5-fbf7-4900-92aa-181b72767e79"). InnerVolumeSpecName "kube-api-access-sllqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.777358 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sllqw\" (UniqueName: \"kubernetes.io/projected/4acec0f5-fbf7-4900-92aa-181b72767e79-kube-api-access-sllqw\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.804655 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4acec0f5-fbf7-4900-92aa-181b72767e79" (UID: "4acec0f5-fbf7-4900-92aa-181b72767e79"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.807008 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-config" (OuterVolumeSpecName: "config") pod "4acec0f5-fbf7-4900-92aa-181b72767e79" (UID: "4acec0f5-fbf7-4900-92aa-181b72767e79"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.819632 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4acec0f5-fbf7-4900-92aa-181b72767e79" (UID: "4acec0f5-fbf7-4900-92aa-181b72767e79"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.834430 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4acec0f5-fbf7-4900-92aa-181b72767e79" (UID: "4acec0f5-fbf7-4900-92aa-181b72767e79"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.862416 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4acec0f5-fbf7-4900-92aa-181b72767e79" (UID: "4acec0f5-fbf7-4900-92aa-181b72767e79"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.880535 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.880567 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.880576 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.880584 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:42 crc kubenswrapper[4696]: I0321 08:51:42.880592 4696 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4acec0f5-fbf7-4900-92aa-181b72767e79-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:43 crc kubenswrapper[4696]: I0321 08:51:43.168702 4696 generic.go:334] "Generic (PLEG): container finished" podID="c9fe9d98-b02d-4fdd-918e-c296d52f7efd" containerID="09b8469c14b96082830c0af00f204b9a0458de32dfe6f4d4e2713421920141bb" exitCode=0 Mar 21 08:51:43 crc kubenswrapper[4696]: I0321 08:51:43.168861 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f64749dc-tqgff" event={"ID":"c9fe9d98-b02d-4fdd-918e-c296d52f7efd","Type":"ContainerDied","Data":"09b8469c14b96082830c0af00f204b9a0458de32dfe6f4d4e2713421920141bb"} Mar 21 08:51:43 crc kubenswrapper[4696]: I0321 08:51:43.169791 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f64749dc-tqgff" event={"ID":"c9fe9d98-b02d-4fdd-918e-c296d52f7efd","Type":"ContainerStarted","Data":"fb4dcfc99d5363b56f8c4007e3f941a4c69369f9cce589e14bfe9e0bf6ddd246"} Mar 21 08:51:43 crc kubenswrapper[4696]: I0321 08:51:43.180206 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-fq9xb" event={"ID":"1f8a2d50-6415-4d3a-8cb4-bbbfce785842","Type":"ContainerStarted","Data":"f88fc9473fb238eb74c3a9d5c3c6794ca0df87320ffa9a7de517cdfb9c6f0d5b"} Mar 21 08:51:43 crc kubenswrapper[4696]: I0321 08:51:43.185768 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" Mar 21 08:51:43 crc kubenswrapper[4696]: I0321 08:51:43.185876 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fd9b586ff-xr8tz" event={"ID":"4acec0f5-fbf7-4900-92aa-181b72767e79","Type":"ContainerDied","Data":"69f6eba6b1c8ebe80e384b5f2ef91f9adfb73a525452be2737c44329e6fbc632"} Mar 21 08:51:43 crc kubenswrapper[4696]: I0321 08:51:43.185924 4696 scope.go:117] "RemoveContainer" containerID="98c8fe9d6eb7efe4cbb38d9718d2575729319445cfe0682e2f960ac47a953850" Mar 21 08:51:43 crc kubenswrapper[4696]: I0321 08:51:43.402477 4696 scope.go:117] "RemoveContainer" containerID="4c7c5acb31fa5b1d43fabd39637850ce34e084b6c8ea6c6ebf1370f8dacaef67" Mar 21 08:51:43 crc kubenswrapper[4696]: I0321 08:51:43.431916 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-db-sync-fq9xb" podStartSLOduration=2.119563888 podStartE2EDuration="37.431898154s" podCreationTimestamp="2026-03-21 08:51:06 +0000 UTC" firstStartedPulling="2026-03-21 08:51:07.496501562 +0000 UTC m=+1401.617382275" lastFinishedPulling="2026-03-21 08:51:42.808835828 +0000 UTC m=+1436.929716541" observedRunningTime="2026-03-21 08:51:43.227074182 +0000 UTC m=+1437.347954915" watchObservedRunningTime="2026-03-21 08:51:43.431898154 +0000 UTC m=+1437.552778867" Mar 21 08:51:43 crc kubenswrapper[4696]: I0321 08:51:43.439256 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5fd9b586ff-xr8tz"] Mar 21 08:51:43 crc kubenswrapper[4696]: I0321 08:51:43.458484 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5fd9b586ff-xr8tz"] Mar 21 08:51:44 crc kubenswrapper[4696]: I0321 08:51:44.198282 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f64749dc-tqgff" event={"ID":"c9fe9d98-b02d-4fdd-918e-c296d52f7efd","Type":"ContainerStarted","Data":"84dc5be5a5d437ed33370aa50c9c23d0f14d64a5fd239c423cc31a71a4e5c638"} Mar 21 08:51:44 crc kubenswrapper[4696]: I0321 08:51:44.198444 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:44 crc kubenswrapper[4696]: I0321 08:51:44.231192 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85f64749dc-tqgff" podStartSLOduration=3.231173199 podStartE2EDuration="3.231173199s" podCreationTimestamp="2026-03-21 08:51:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:51:44.223304408 +0000 UTC m=+1438.344185121" watchObservedRunningTime="2026-03-21 08:51:44.231173199 +0000 UTC m=+1438.352053902" Mar 21 08:51:44 crc kubenswrapper[4696]: I0321 08:51:44.548130 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4acec0f5-fbf7-4900-92aa-181b72767e79" path="/var/lib/kubelet/pods/4acec0f5-fbf7-4900-92aa-181b72767e79/volumes" Mar 21 08:51:46 crc kubenswrapper[4696]: I0321 08:51:46.240368 4696 generic.go:334] "Generic (PLEG): container finished" podID="1f8a2d50-6415-4d3a-8cb4-bbbfce785842" containerID="f88fc9473fb238eb74c3a9d5c3c6794ca0df87320ffa9a7de517cdfb9c6f0d5b" exitCode=0 Mar 21 08:51:46 crc kubenswrapper[4696]: I0321 08:51:46.240807 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-fq9xb" event={"ID":"1f8a2d50-6415-4d3a-8cb4-bbbfce785842","Type":"ContainerDied","Data":"f88fc9473fb238eb74c3a9d5c3c6794ca0df87320ffa9a7de517cdfb9c6f0d5b"} Mar 21 08:51:47 crc kubenswrapper[4696]: I0321 08:51:47.011835 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-br9c4" Mar 21 08:51:47 crc kubenswrapper[4696]: I0321 08:51:47.080799 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-br9c4" Mar 21 08:51:47 crc kubenswrapper[4696]: I0321 08:51:47.248262 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-br9c4"] Mar 21 08:51:47 crc kubenswrapper[4696]: I0321 08:51:47.891427 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-fq9xb" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.000740 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-config-data\") pod \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\" (UID: \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\") " Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.000902 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kzqv\" (UniqueName: \"kubernetes.io/projected/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-kube-api-access-5kzqv\") pod \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\" (UID: \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\") " Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.001003 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-combined-ca-bundle\") pod \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\" (UID: \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\") " Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.001036 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-certs\") pod \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\" (UID: \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\") " Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.001067 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-scripts\") pod \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\" (UID: \"1f8a2d50-6415-4d3a-8cb4-bbbfce785842\") " Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.007873 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-kube-api-access-5kzqv" (OuterVolumeSpecName: "kube-api-access-5kzqv") pod "1f8a2d50-6415-4d3a-8cb4-bbbfce785842" (UID: "1f8a2d50-6415-4d3a-8cb4-bbbfce785842"). InnerVolumeSpecName "kube-api-access-5kzqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.008945 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-scripts" (OuterVolumeSpecName: "scripts") pod "1f8a2d50-6415-4d3a-8cb4-bbbfce785842" (UID: "1f8a2d50-6415-4d3a-8cb4-bbbfce785842"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.012393 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-certs" (OuterVolumeSpecName: "certs") pod "1f8a2d50-6415-4d3a-8cb4-bbbfce785842" (UID: "1f8a2d50-6415-4d3a-8cb4-bbbfce785842"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.048227 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1f8a2d50-6415-4d3a-8cb4-bbbfce785842" (UID: "1f8a2d50-6415-4d3a-8cb4-bbbfce785842"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.048843 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-config-data" (OuterVolumeSpecName: "config-data") pod "1f8a2d50-6415-4d3a-8cb4-bbbfce785842" (UID: "1f8a2d50-6415-4d3a-8cb4-bbbfce785842"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.103803 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.103857 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.103871 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kzqv\" (UniqueName: \"kubernetes.io/projected/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-kube-api-access-5kzqv\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.103883 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.103892 4696 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/1f8a2d50-6415-4d3a-8cb4-bbbfce785842-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.260977 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-fq9xb" event={"ID":"1f8a2d50-6415-4d3a-8cb4-bbbfce785842","Type":"ContainerDied","Data":"d1d6775ac797c68b94bd80ce81230d525dd64dd919c8958f29709727cbef35a3"} Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.261047 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1d6775ac797c68b94bd80ce81230d525dd64dd919c8958f29709727cbef35a3" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.261164 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-br9c4" podUID="817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a" containerName="registry-server" containerID="cri-o://f1f42ab54713d20c94a91d51039358cfd6b2751fbb78b7b016e31b7602c81ee0" gracePeriod=2 Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.261006 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-fq9xb" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.416961 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-storageinit-8r5xw"] Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.429955 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-storageinit-8r5xw"] Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.562158 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d73897de-7ac5-4991-aabf-b36cc6ef424d" path="/var/lib/kubelet/pods/d73897de-7ac5-4991-aabf-b36cc6ef424d/volumes" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.572662 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-storageinit-8r5zz"] Mar 21 08:51:48 crc kubenswrapper[4696]: E0321 08:51:48.578337 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f8a2d50-6415-4d3a-8cb4-bbbfce785842" containerName="cloudkitty-db-sync" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.578370 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f8a2d50-6415-4d3a-8cb4-bbbfce785842" containerName="cloudkitty-db-sync" Mar 21 08:51:48 crc kubenswrapper[4696]: E0321 08:51:48.578392 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4acec0f5-fbf7-4900-92aa-181b72767e79" containerName="dnsmasq-dns" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.578400 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4acec0f5-fbf7-4900-92aa-181b72767e79" containerName="dnsmasq-dns" Mar 21 08:51:48 crc kubenswrapper[4696]: E0321 08:51:48.578425 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4acec0f5-fbf7-4900-92aa-181b72767e79" containerName="init" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.578432 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4acec0f5-fbf7-4900-92aa-181b72767e79" containerName="init" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.578684 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="4acec0f5-fbf7-4900-92aa-181b72767e79" containerName="dnsmasq-dns" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.578723 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f8a2d50-6415-4d3a-8cb4-bbbfce785842" containerName="cloudkitty-db-sync" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.579664 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-8r5zz" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.592170 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.607902 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-8r5zz"] Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.613644 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e75fe987-8f3c-4f76-9546-c6ab537af3ce-scripts\") pod \"cloudkitty-storageinit-8r5zz\" (UID: \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\") " pod="openstack/cloudkitty-storageinit-8r5zz" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.613697 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trkdl\" (UniqueName: \"kubernetes.io/projected/e75fe987-8f3c-4f76-9546-c6ab537af3ce-kube-api-access-trkdl\") pod \"cloudkitty-storageinit-8r5zz\" (UID: \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\") " pod="openstack/cloudkitty-storageinit-8r5zz" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.613856 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fe987-8f3c-4f76-9546-c6ab537af3ce-combined-ca-bundle\") pod \"cloudkitty-storageinit-8r5zz\" (UID: \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\") " pod="openstack/cloudkitty-storageinit-8r5zz" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.613888 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e75fe987-8f3c-4f76-9546-c6ab537af3ce-config-data\") pod \"cloudkitty-storageinit-8r5zz\" (UID: \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\") " pod="openstack/cloudkitty-storageinit-8r5zz" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.614026 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/e75fe987-8f3c-4f76-9546-c6ab537af3ce-certs\") pod \"cloudkitty-storageinit-8r5zz\" (UID: \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\") " pod="openstack/cloudkitty-storageinit-8r5zz" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.715679 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fe987-8f3c-4f76-9546-c6ab537af3ce-combined-ca-bundle\") pod \"cloudkitty-storageinit-8r5zz\" (UID: \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\") " pod="openstack/cloudkitty-storageinit-8r5zz" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.715741 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e75fe987-8f3c-4f76-9546-c6ab537af3ce-config-data\") pod \"cloudkitty-storageinit-8r5zz\" (UID: \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\") " pod="openstack/cloudkitty-storageinit-8r5zz" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.715831 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/e75fe987-8f3c-4f76-9546-c6ab537af3ce-certs\") pod \"cloudkitty-storageinit-8r5zz\" (UID: \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\") " pod="openstack/cloudkitty-storageinit-8r5zz" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.715904 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e75fe987-8f3c-4f76-9546-c6ab537af3ce-scripts\") pod \"cloudkitty-storageinit-8r5zz\" (UID: \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\") " pod="openstack/cloudkitty-storageinit-8r5zz" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.715939 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trkdl\" (UniqueName: \"kubernetes.io/projected/e75fe987-8f3c-4f76-9546-c6ab537af3ce-kube-api-access-trkdl\") pod \"cloudkitty-storageinit-8r5zz\" (UID: \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\") " pod="openstack/cloudkitty-storageinit-8r5zz" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.720302 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fe987-8f3c-4f76-9546-c6ab537af3ce-combined-ca-bundle\") pod \"cloudkitty-storageinit-8r5zz\" (UID: \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\") " pod="openstack/cloudkitty-storageinit-8r5zz" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.721716 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e75fe987-8f3c-4f76-9546-c6ab537af3ce-scripts\") pod \"cloudkitty-storageinit-8r5zz\" (UID: \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\") " pod="openstack/cloudkitty-storageinit-8r5zz" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.722656 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e75fe987-8f3c-4f76-9546-c6ab537af3ce-config-data\") pod \"cloudkitty-storageinit-8r5zz\" (UID: \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\") " pod="openstack/cloudkitty-storageinit-8r5zz" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.731865 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/e75fe987-8f3c-4f76-9546-c6ab537af3ce-certs\") pod \"cloudkitty-storageinit-8r5zz\" (UID: \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\") " pod="openstack/cloudkitty-storageinit-8r5zz" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.740402 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trkdl\" (UniqueName: \"kubernetes.io/projected/e75fe987-8f3c-4f76-9546-c6ab537af3ce-kube-api-access-trkdl\") pod \"cloudkitty-storageinit-8r5zz\" (UID: \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\") " pod="openstack/cloudkitty-storageinit-8r5zz" Mar 21 08:51:48 crc kubenswrapper[4696]: I0321 08:51:48.923213 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-8r5zz" Mar 21 08:51:49 crc kubenswrapper[4696]: I0321 08:51:49.304411 4696 generic.go:334] "Generic (PLEG): container finished" podID="817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a" containerID="f1f42ab54713d20c94a91d51039358cfd6b2751fbb78b7b016e31b7602c81ee0" exitCode=0 Mar 21 08:51:49 crc kubenswrapper[4696]: I0321 08:51:49.304756 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-br9c4" event={"ID":"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a","Type":"ContainerDied","Data":"f1f42ab54713d20c94a91d51039358cfd6b2751fbb78b7b016e31b7602c81ee0"} Mar 21 08:51:49 crc kubenswrapper[4696]: I0321 08:51:49.456987 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-br9c4" Mar 21 08:51:49 crc kubenswrapper[4696]: I0321 08:51:49.529081 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-8r5zz"] Mar 21 08:51:49 crc kubenswrapper[4696]: I0321 08:51:49.531225 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a-catalog-content\") pod \"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a\" (UID: \"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a\") " Mar 21 08:51:49 crc kubenswrapper[4696]: I0321 08:51:49.531322 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfntv\" (UniqueName: \"kubernetes.io/projected/817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a-kube-api-access-bfntv\") pod \"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a\" (UID: \"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a\") " Mar 21 08:51:49 crc kubenswrapper[4696]: I0321 08:51:49.531342 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a-utilities\") pod \"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a\" (UID: \"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a\") " Mar 21 08:51:49 crc kubenswrapper[4696]: I0321 08:51:49.532972 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a-utilities" (OuterVolumeSpecName: "utilities") pod "817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a" (UID: "817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:51:49 crc kubenswrapper[4696]: I0321 08:51:49.538455 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a-kube-api-access-bfntv" (OuterVolumeSpecName: "kube-api-access-bfntv") pod "817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a" (UID: "817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a"). InnerVolumeSpecName "kube-api-access-bfntv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:51:49 crc kubenswrapper[4696]: I0321 08:51:49.636114 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfntv\" (UniqueName: \"kubernetes.io/projected/817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a-kube-api-access-bfntv\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:49 crc kubenswrapper[4696]: I0321 08:51:49.636164 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:49 crc kubenswrapper[4696]: I0321 08:51:49.713502 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a" (UID: "817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:51:49 crc kubenswrapper[4696]: I0321 08:51:49.738146 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:50 crc kubenswrapper[4696]: I0321 08:51:50.317613 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-8r5zz" event={"ID":"e75fe987-8f3c-4f76-9546-c6ab537af3ce","Type":"ContainerStarted","Data":"5137937cfb80b886d262e80869076dfafade90bbe4c64c139378725b1d8e437d"} Mar 21 08:51:50 crc kubenswrapper[4696]: I0321 08:51:50.317919 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-8r5zz" event={"ID":"e75fe987-8f3c-4f76-9546-c6ab537af3ce","Type":"ContainerStarted","Data":"6976bd0bf87b65d208005cfd0ef8c1529020345555ae341768797925cfcc48e9"} Mar 21 08:51:50 crc kubenswrapper[4696]: I0321 08:51:50.321855 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-br9c4" event={"ID":"817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a","Type":"ContainerDied","Data":"8b8a41951178c42396c7ecdfd19805743ec595d6378554a8f7008c7b82fc49c9"} Mar 21 08:51:50 crc kubenswrapper[4696]: I0321 08:51:50.321912 4696 scope.go:117] "RemoveContainer" containerID="f1f42ab54713d20c94a91d51039358cfd6b2751fbb78b7b016e31b7602c81ee0" Mar 21 08:51:50 crc kubenswrapper[4696]: I0321 08:51:50.321909 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-br9c4" Mar 21 08:51:50 crc kubenswrapper[4696]: I0321 08:51:50.338481 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-storageinit-8r5zz" podStartSLOduration=2.338459688 podStartE2EDuration="2.338459688s" podCreationTimestamp="2026-03-21 08:51:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:51:50.335188117 +0000 UTC m=+1444.456068830" watchObservedRunningTime="2026-03-21 08:51:50.338459688 +0000 UTC m=+1444.459340401" Mar 21 08:51:50 crc kubenswrapper[4696]: I0321 08:51:50.363552 4696 scope.go:117] "RemoveContainer" containerID="3217cfe47d98f53b09d21bbc10fad865144d753bdccf955cd24aef22067aab64" Mar 21 08:51:50 crc kubenswrapper[4696]: I0321 08:51:50.379480 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-br9c4"] Mar 21 08:51:50 crc kubenswrapper[4696]: I0321 08:51:50.406518 4696 scope.go:117] "RemoveContainer" containerID="58e01952ce0256947c0ae7181c811bc2bd35b1555b6ba25a6b7f5d438f2726f1" Mar 21 08:51:50 crc kubenswrapper[4696]: I0321 08:51:50.409158 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-br9c4"] Mar 21 08:51:50 crc kubenswrapper[4696]: I0321 08:51:50.548750 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a" path="/var/lib/kubelet/pods/817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a/volumes" Mar 21 08:51:51 crc kubenswrapper[4696]: I0321 08:51:51.971966 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85f64749dc-tqgff" Mar 21 08:51:52 crc kubenswrapper[4696]: I0321 08:51:52.039033 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dbb88bf8c-gjwx5"] Mar 21 08:51:52 crc kubenswrapper[4696]: I0321 08:51:52.039314 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" podUID="1e77a97c-92ea-4bc5-bb15-b01d360220b6" containerName="dnsmasq-dns" containerID="cri-o://6901cb5ddec0972f962e20bf62f5fbd48b07096d1d54cf8ac35af0ead3781b8b" gracePeriod=10 Mar 21 08:51:52 crc kubenswrapper[4696]: I0321 08:51:52.363578 4696 generic.go:334] "Generic (PLEG): container finished" podID="1e77a97c-92ea-4bc5-bb15-b01d360220b6" containerID="6901cb5ddec0972f962e20bf62f5fbd48b07096d1d54cf8ac35af0ead3781b8b" exitCode=0 Mar 21 08:51:52 crc kubenswrapper[4696]: I0321 08:51:52.363987 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" event={"ID":"1e77a97c-92ea-4bc5-bb15-b01d360220b6","Type":"ContainerDied","Data":"6901cb5ddec0972f962e20bf62f5fbd48b07096d1d54cf8ac35af0ead3781b8b"} Mar 21 08:51:52 crc kubenswrapper[4696]: I0321 08:51:52.366007 4696 generic.go:334] "Generic (PLEG): container finished" podID="e75fe987-8f3c-4f76-9546-c6ab537af3ce" containerID="5137937cfb80b886d262e80869076dfafade90bbe4c64c139378725b1d8e437d" exitCode=0 Mar 21 08:51:52 crc kubenswrapper[4696]: I0321 08:51:52.366039 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-8r5zz" event={"ID":"e75fe987-8f3c-4f76-9546-c6ab537af3ce","Type":"ContainerDied","Data":"5137937cfb80b886d262e80869076dfafade90bbe4c64c139378725b1d8e437d"} Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.172178 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.205412 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-ovsdbserver-sb\") pod \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.205475 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-dns-swift-storage-0\") pod \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.205731 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkhbf\" (UniqueName: \"kubernetes.io/projected/1e77a97c-92ea-4bc5-bb15-b01d360220b6-kube-api-access-lkhbf\") pod \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.205811 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-config\") pod \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.205858 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-dns-svc\") pod \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.205891 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-openstack-edpm-ipam\") pod \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.205965 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-ovsdbserver-nb\") pod \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\" (UID: \"1e77a97c-92ea-4bc5-bb15-b01d360220b6\") " Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.259020 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e77a97c-92ea-4bc5-bb15-b01d360220b6-kube-api-access-lkhbf" (OuterVolumeSpecName: "kube-api-access-lkhbf") pod "1e77a97c-92ea-4bc5-bb15-b01d360220b6" (UID: "1e77a97c-92ea-4bc5-bb15-b01d360220b6"). InnerVolumeSpecName "kube-api-access-lkhbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.310597 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkhbf\" (UniqueName: \"kubernetes.io/projected/1e77a97c-92ea-4bc5-bb15-b01d360220b6-kube-api-access-lkhbf\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.340705 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1e77a97c-92ea-4bc5-bb15-b01d360220b6" (UID: "1e77a97c-92ea-4bc5-bb15-b01d360220b6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.347285 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-config" (OuterVolumeSpecName: "config") pod "1e77a97c-92ea-4bc5-bb15-b01d360220b6" (UID: "1e77a97c-92ea-4bc5-bb15-b01d360220b6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.348139 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1e77a97c-92ea-4bc5-bb15-b01d360220b6" (UID: "1e77a97c-92ea-4bc5-bb15-b01d360220b6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.362336 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "1e77a97c-92ea-4bc5-bb15-b01d360220b6" (UID: "1e77a97c-92ea-4bc5-bb15-b01d360220b6"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.380893 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.381659 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1e77a97c-92ea-4bc5-bb15-b01d360220b6" (UID: "1e77a97c-92ea-4bc5-bb15-b01d360220b6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.381745 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbb88bf8c-gjwx5" event={"ID":"1e77a97c-92ea-4bc5-bb15-b01d360220b6","Type":"ContainerDied","Data":"cebca7f8cba161f8e5e5e0bff62e1a7d4c22d953179115d57357f890b77aa577"} Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.381786 4696 scope.go:117] "RemoveContainer" containerID="6901cb5ddec0972f962e20bf62f5fbd48b07096d1d54cf8ac35af0ead3781b8b" Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.390153 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1e77a97c-92ea-4bc5-bb15-b01d360220b6" (UID: "1e77a97c-92ea-4bc5-bb15-b01d360220b6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.415451 4696 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-config\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.415484 4696 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.415496 4696 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.415507 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.415542 4696 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.415553 4696 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1e77a97c-92ea-4bc5-bb15-b01d360220b6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.479652 4696 scope.go:117] "RemoveContainer" containerID="56e2fc9386018f8d98cb4e8c2eedfd1c0d6cd7f88c2355450d0bc4edac6c3379" Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.719324 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dbb88bf8c-gjwx5"] Mar 21 08:51:53 crc kubenswrapper[4696]: I0321 08:51:53.730659 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-dbb88bf8c-gjwx5"] Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.109846 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-8r5zz" Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.131153 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fe987-8f3c-4f76-9546-c6ab537af3ce-combined-ca-bundle\") pod \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\" (UID: \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\") " Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.131378 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trkdl\" (UniqueName: \"kubernetes.io/projected/e75fe987-8f3c-4f76-9546-c6ab537af3ce-kube-api-access-trkdl\") pod \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\" (UID: \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\") " Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.131503 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e75fe987-8f3c-4f76-9546-c6ab537af3ce-scripts\") pod \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\" (UID: \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\") " Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.131549 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e75fe987-8f3c-4f76-9546-c6ab537af3ce-config-data\") pod \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\" (UID: \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\") " Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.131578 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/e75fe987-8f3c-4f76-9546-c6ab537af3ce-certs\") pod \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\" (UID: \"e75fe987-8f3c-4f76-9546-c6ab537af3ce\") " Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.136346 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e75fe987-8f3c-4f76-9546-c6ab537af3ce-certs" (OuterVolumeSpecName: "certs") pod "e75fe987-8f3c-4f76-9546-c6ab537af3ce" (UID: "e75fe987-8f3c-4f76-9546-c6ab537af3ce"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.136657 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e75fe987-8f3c-4f76-9546-c6ab537af3ce-kube-api-access-trkdl" (OuterVolumeSpecName: "kube-api-access-trkdl") pod "e75fe987-8f3c-4f76-9546-c6ab537af3ce" (UID: "e75fe987-8f3c-4f76-9546-c6ab537af3ce"). InnerVolumeSpecName "kube-api-access-trkdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.136654 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e75fe987-8f3c-4f76-9546-c6ab537af3ce-scripts" (OuterVolumeSpecName: "scripts") pod "e75fe987-8f3c-4f76-9546-c6ab537af3ce" (UID: "e75fe987-8f3c-4f76-9546-c6ab537af3ce"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.167086 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e75fe987-8f3c-4f76-9546-c6ab537af3ce-config-data" (OuterVolumeSpecName: "config-data") pod "e75fe987-8f3c-4f76-9546-c6ab537af3ce" (UID: "e75fe987-8f3c-4f76-9546-c6ab537af3ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.168492 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e75fe987-8f3c-4f76-9546-c6ab537af3ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e75fe987-8f3c-4f76-9546-c6ab537af3ce" (UID: "e75fe987-8f3c-4f76-9546-c6ab537af3ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.234305 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trkdl\" (UniqueName: \"kubernetes.io/projected/e75fe987-8f3c-4f76-9546-c6ab537af3ce-kube-api-access-trkdl\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.234346 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e75fe987-8f3c-4f76-9546-c6ab537af3ce-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.234358 4696 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/e75fe987-8f3c-4f76-9546-c6ab537af3ce-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.234368 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e75fe987-8f3c-4f76-9546-c6ab537af3ce-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.234379 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75fe987-8f3c-4f76-9546-c6ab537af3ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.395341 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-8r5zz" event={"ID":"e75fe987-8f3c-4f76-9546-c6ab537af3ce","Type":"ContainerDied","Data":"6976bd0bf87b65d208005cfd0ef8c1529020345555ae341768797925cfcc48e9"} Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.395377 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6976bd0bf87b65d208005cfd0ef8c1529020345555ae341768797925cfcc48e9" Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.395429 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-8r5zz" Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.503359 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.503656 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="d623e522-1d67-45ef-8ebd-a8a37e12f045" containerName="cloudkitty-api" containerID="cri-o://68c5a483e2f9a4917af42f3a50fcac8a07ccc00aa1a9b31b189b930f3f5d071c" gracePeriod=30 Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.503626 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="d623e522-1d67-45ef-8ebd-a8a37e12f045" containerName="cloudkitty-api-log" containerID="cri-o://122f960a9aa1d8bd0345a90c369c67c7c94fa1b94e8fcf6019667927ee380e6a" gracePeriod=30 Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.521638 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.522454 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-proc-0" podUID="623a2b50-8a77-47df-800c-5ec7c237255c" containerName="cloudkitty-proc" containerID="cri-o://b1600882e3e31cebfdf37e53812fcf328b05af85029dd298a9919826dcf85349" gracePeriod=30 Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.548073 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e77a97c-92ea-4bc5-bb15-b01d360220b6" path="/var/lib/kubelet/pods/1e77a97c-92ea-4bc5-bb15-b01d360220b6/volumes" Mar 21 08:51:54 crc kubenswrapper[4696]: I0321 08:51:54.621952 4696 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="1e6b9c7f-6556-4ab0-ac07-4c11a662000a" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.238:3000/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 21 08:51:55 crc kubenswrapper[4696]: E0321 08:51:55.186360 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod623a2b50_8a77_47df_800c_5ec7c237255c.slice/crio-conmon-b1600882e3e31cebfdf37e53812fcf328b05af85029dd298a9919826dcf85349.scope\": RecentStats: unable to find data in memory cache]" Mar 21 08:51:55 crc kubenswrapper[4696]: I0321 08:51:55.412722 4696 generic.go:334] "Generic (PLEG): container finished" podID="623a2b50-8a77-47df-800c-5ec7c237255c" containerID="b1600882e3e31cebfdf37e53812fcf328b05af85029dd298a9919826dcf85349" exitCode=0 Mar 21 08:51:55 crc kubenswrapper[4696]: I0321 08:51:55.413110 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"623a2b50-8a77-47df-800c-5ec7c237255c","Type":"ContainerDied","Data":"b1600882e3e31cebfdf37e53812fcf328b05af85029dd298a9919826dcf85349"} Mar 21 08:51:55 crc kubenswrapper[4696]: I0321 08:51:55.436171 4696 generic.go:334] "Generic (PLEG): container finished" podID="d623e522-1d67-45ef-8ebd-a8a37e12f045" containerID="122f960a9aa1d8bd0345a90c369c67c7c94fa1b94e8fcf6019667927ee380e6a" exitCode=143 Mar 21 08:51:55 crc kubenswrapper[4696]: I0321 08:51:55.436227 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"d623e522-1d67-45ef-8ebd-a8a37e12f045","Type":"ContainerDied","Data":"122f960a9aa1d8bd0345a90c369c67c7c94fa1b94e8fcf6019667927ee380e6a"} Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.195007 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.226781 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.289305 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-scripts\") pod \"d623e522-1d67-45ef-8ebd-a8a37e12f045\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.289443 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crcb9\" (UniqueName: \"kubernetes.io/projected/d623e522-1d67-45ef-8ebd-a8a37e12f045-kube-api-access-crcb9\") pod \"d623e522-1d67-45ef-8ebd-a8a37e12f045\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.289465 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-public-tls-certs\") pod \"d623e522-1d67-45ef-8ebd-a8a37e12f045\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.289495 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d623e522-1d67-45ef-8ebd-a8a37e12f045-certs\") pod \"d623e522-1d67-45ef-8ebd-a8a37e12f045\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.289534 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-config-data-custom\") pod \"d623e522-1d67-45ef-8ebd-a8a37e12f045\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.289568 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-config-data\") pod \"d623e522-1d67-45ef-8ebd-a8a37e12f045\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.289657 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-internal-tls-certs\") pod \"d623e522-1d67-45ef-8ebd-a8a37e12f045\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.289703 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d623e522-1d67-45ef-8ebd-a8a37e12f045-logs\") pod \"d623e522-1d67-45ef-8ebd-a8a37e12f045\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.289810 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-combined-ca-bundle\") pod \"d623e522-1d67-45ef-8ebd-a8a37e12f045\" (UID: \"d623e522-1d67-45ef-8ebd-a8a37e12f045\") " Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.292602 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d623e522-1d67-45ef-8ebd-a8a37e12f045-logs" (OuterVolumeSpecName: "logs") pod "d623e522-1d67-45ef-8ebd-a8a37e12f045" (UID: "d623e522-1d67-45ef-8ebd-a8a37e12f045"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.297899 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d623e522-1d67-45ef-8ebd-a8a37e12f045" (UID: "d623e522-1d67-45ef-8ebd-a8a37e12f045"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.300203 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d623e522-1d67-45ef-8ebd-a8a37e12f045-certs" (OuterVolumeSpecName: "certs") pod "d623e522-1d67-45ef-8ebd-a8a37e12f045" (UID: "d623e522-1d67-45ef-8ebd-a8a37e12f045"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.315777 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-scripts" (OuterVolumeSpecName: "scripts") pod "d623e522-1d67-45ef-8ebd-a8a37e12f045" (UID: "d623e522-1d67-45ef-8ebd-a8a37e12f045"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.321002 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d623e522-1d67-45ef-8ebd-a8a37e12f045-kube-api-access-crcb9" (OuterVolumeSpecName: "kube-api-access-crcb9") pod "d623e522-1d67-45ef-8ebd-a8a37e12f045" (UID: "d623e522-1d67-45ef-8ebd-a8a37e12f045"). InnerVolumeSpecName "kube-api-access-crcb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.328928 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-config-data" (OuterVolumeSpecName: "config-data") pod "d623e522-1d67-45ef-8ebd-a8a37e12f045" (UID: "d623e522-1d67-45ef-8ebd-a8a37e12f045"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.336898 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.361842 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d623e522-1d67-45ef-8ebd-a8a37e12f045" (UID: "d623e522-1d67-45ef-8ebd-a8a37e12f045"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.391002 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-config-data\") pod \"623a2b50-8a77-47df-800c-5ec7c237255c\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.391200 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/623a2b50-8a77-47df-800c-5ec7c237255c-certs\") pod \"623a2b50-8a77-47df-800c-5ec7c237255c\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.391240 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-config-data-custom\") pod \"623a2b50-8a77-47df-800c-5ec7c237255c\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.391333 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fs2sr\" (UniqueName: \"kubernetes.io/projected/623a2b50-8a77-47df-800c-5ec7c237255c-kube-api-access-fs2sr\") pod \"623a2b50-8a77-47df-800c-5ec7c237255c\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.391373 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-combined-ca-bundle\") pod \"623a2b50-8a77-47df-800c-5ec7c237255c\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.391428 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-scripts\") pod \"623a2b50-8a77-47df-800c-5ec7c237255c\" (UID: \"623a2b50-8a77-47df-800c-5ec7c237255c\") " Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.391857 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.391874 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.391884 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crcb9\" (UniqueName: \"kubernetes.io/projected/d623e522-1d67-45ef-8ebd-a8a37e12f045-kube-api-access-crcb9\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.391894 4696 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d623e522-1d67-45ef-8ebd-a8a37e12f045-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.391902 4696 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.391910 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.391919 4696 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d623e522-1d67-45ef-8ebd-a8a37e12f045-logs\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.395198 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-scripts" (OuterVolumeSpecName: "scripts") pod "623a2b50-8a77-47df-800c-5ec7c237255c" (UID: "623a2b50-8a77-47df-800c-5ec7c237255c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.396639 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "623a2b50-8a77-47df-800c-5ec7c237255c" (UID: "623a2b50-8a77-47df-800c-5ec7c237255c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.397957 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/623a2b50-8a77-47df-800c-5ec7c237255c-certs" (OuterVolumeSpecName: "certs") pod "623a2b50-8a77-47df-800c-5ec7c237255c" (UID: "623a2b50-8a77-47df-800c-5ec7c237255c"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.399400 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/623a2b50-8a77-47df-800c-5ec7c237255c-kube-api-access-fs2sr" (OuterVolumeSpecName: "kube-api-access-fs2sr") pod "623a2b50-8a77-47df-800c-5ec7c237255c" (UID: "623a2b50-8a77-47df-800c-5ec7c237255c"). InnerVolumeSpecName "kube-api-access-fs2sr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.403358 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d623e522-1d67-45ef-8ebd-a8a37e12f045" (UID: "d623e522-1d67-45ef-8ebd-a8a37e12f045"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.406201 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d623e522-1d67-45ef-8ebd-a8a37e12f045" (UID: "d623e522-1d67-45ef-8ebd-a8a37e12f045"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.424856 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-config-data" (OuterVolumeSpecName: "config-data") pod "623a2b50-8a77-47df-800c-5ec7c237255c" (UID: "623a2b50-8a77-47df-800c-5ec7c237255c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.435474 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "623a2b50-8a77-47df-800c-5ec7c237255c" (UID: "623a2b50-8a77-47df-800c-5ec7c237255c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.447173 4696 generic.go:334] "Generic (PLEG): container finished" podID="d623e522-1d67-45ef-8ebd-a8a37e12f045" containerID="68c5a483e2f9a4917af42f3a50fcac8a07ccc00aa1a9b31b189b930f3f5d071c" exitCode=0 Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.447231 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"d623e522-1d67-45ef-8ebd-a8a37e12f045","Type":"ContainerDied","Data":"68c5a483e2f9a4917af42f3a50fcac8a07ccc00aa1a9b31b189b930f3f5d071c"} Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.447257 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"d623e522-1d67-45ef-8ebd-a8a37e12f045","Type":"ContainerDied","Data":"d55f509a352fa8a26feffdf52136d37fceac48489792434ec6dbf155bca37171"} Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.447272 4696 scope.go:117] "RemoveContainer" containerID="68c5a483e2f9a4917af42f3a50fcac8a07ccc00aa1a9b31b189b930f3f5d071c" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.447384 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.451527 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"623a2b50-8a77-47df-800c-5ec7c237255c","Type":"ContainerDied","Data":"7368dbc952d8a83cdc4e59b462d771cda60acfd48762b5f9422ce41553ed3d96"} Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.451601 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.488307 4696 scope.go:117] "RemoveContainer" containerID="122f960a9aa1d8bd0345a90c369c67c7c94fa1b94e8fcf6019667927ee380e6a" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.494031 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fs2sr\" (UniqueName: \"kubernetes.io/projected/623a2b50-8a77-47df-800c-5ec7c237255c-kube-api-access-fs2sr\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.494071 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.494081 4696 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-scripts\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.494090 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.494099 4696 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.494107 4696 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/623a2b50-8a77-47df-800c-5ec7c237255c-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.494115 4696 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/623a2b50-8a77-47df-800c-5ec7c237255c-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.494124 4696 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d623e522-1d67-45ef-8ebd-a8a37e12f045-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.501890 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.540080 4696 scope.go:117] "RemoveContainer" containerID="68c5a483e2f9a4917af42f3a50fcac8a07ccc00aa1a9b31b189b930f3f5d071c" Mar 21 08:51:56 crc kubenswrapper[4696]: E0321 08:51:56.540443 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68c5a483e2f9a4917af42f3a50fcac8a07ccc00aa1a9b31b189b930f3f5d071c\": container with ID starting with 68c5a483e2f9a4917af42f3a50fcac8a07ccc00aa1a9b31b189b930f3f5d071c not found: ID does not exist" containerID="68c5a483e2f9a4917af42f3a50fcac8a07ccc00aa1a9b31b189b930f3f5d071c" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.540474 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68c5a483e2f9a4917af42f3a50fcac8a07ccc00aa1a9b31b189b930f3f5d071c"} err="failed to get container status \"68c5a483e2f9a4917af42f3a50fcac8a07ccc00aa1a9b31b189b930f3f5d071c\": rpc error: code = NotFound desc = could not find container \"68c5a483e2f9a4917af42f3a50fcac8a07ccc00aa1a9b31b189b930f3f5d071c\": container with ID starting with 68c5a483e2f9a4917af42f3a50fcac8a07ccc00aa1a9b31b189b930f3f5d071c not found: ID does not exist" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.540494 4696 scope.go:117] "RemoveContainer" containerID="122f960a9aa1d8bd0345a90c369c67c7c94fa1b94e8fcf6019667927ee380e6a" Mar 21 08:51:56 crc kubenswrapper[4696]: E0321 08:51:56.540736 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"122f960a9aa1d8bd0345a90c369c67c7c94fa1b94e8fcf6019667927ee380e6a\": container with ID starting with 122f960a9aa1d8bd0345a90c369c67c7c94fa1b94e8fcf6019667927ee380e6a not found: ID does not exist" containerID="122f960a9aa1d8bd0345a90c369c67c7c94fa1b94e8fcf6019667927ee380e6a" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.540755 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"122f960a9aa1d8bd0345a90c369c67c7c94fa1b94e8fcf6019667927ee380e6a"} err="failed to get container status \"122f960a9aa1d8bd0345a90c369c67c7c94fa1b94e8fcf6019667927ee380e6a\": rpc error: code = NotFound desc = could not find container \"122f960a9aa1d8bd0345a90c369c67c7c94fa1b94e8fcf6019667927ee380e6a\": container with ID starting with 122f960a9aa1d8bd0345a90c369c67c7c94fa1b94e8fcf6019667927ee380e6a not found: ID does not exist" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.540766 4696 scope.go:117] "RemoveContainer" containerID="b1600882e3e31cebfdf37e53812fcf328b05af85029dd298a9919826dcf85349" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.547419 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-api-0"] Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.561276 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d623e522-1d67-45ef-8ebd-a8a37e12f045" path="/var/lib/kubelet/pods/d623e522-1d67-45ef-8ebd-a8a37e12f045/volumes" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.562112 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.576734 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-proc-0"] Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.591228 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-api-0"] Mar 21 08:51:56 crc kubenswrapper[4696]: E0321 08:51:56.591846 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a" containerName="extract-content" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.591867 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a" containerName="extract-content" Mar 21 08:51:56 crc kubenswrapper[4696]: E0321 08:51:56.591892 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a" containerName="registry-server" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.591899 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a" containerName="registry-server" Mar 21 08:51:56 crc kubenswrapper[4696]: E0321 08:51:56.591927 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e77a97c-92ea-4bc5-bb15-b01d360220b6" containerName="init" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.591933 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e77a97c-92ea-4bc5-bb15-b01d360220b6" containerName="init" Mar 21 08:51:56 crc kubenswrapper[4696]: E0321 08:51:56.591943 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d623e522-1d67-45ef-8ebd-a8a37e12f045" containerName="cloudkitty-api-log" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.591949 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d623e522-1d67-45ef-8ebd-a8a37e12f045" containerName="cloudkitty-api-log" Mar 21 08:51:56 crc kubenswrapper[4696]: E0321 08:51:56.591962 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e75fe987-8f3c-4f76-9546-c6ab537af3ce" containerName="cloudkitty-storageinit" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.591969 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e75fe987-8f3c-4f76-9546-c6ab537af3ce" containerName="cloudkitty-storageinit" Mar 21 08:51:56 crc kubenswrapper[4696]: E0321 08:51:56.591980 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="623a2b50-8a77-47df-800c-5ec7c237255c" containerName="cloudkitty-proc" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.591986 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="623a2b50-8a77-47df-800c-5ec7c237255c" containerName="cloudkitty-proc" Mar 21 08:51:56 crc kubenswrapper[4696]: E0321 08:51:56.592005 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e77a97c-92ea-4bc5-bb15-b01d360220b6" containerName="dnsmasq-dns" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.592010 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e77a97c-92ea-4bc5-bb15-b01d360220b6" containerName="dnsmasq-dns" Mar 21 08:51:56 crc kubenswrapper[4696]: E0321 08:51:56.592027 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a" containerName="extract-utilities" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.592032 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a" containerName="extract-utilities" Mar 21 08:51:56 crc kubenswrapper[4696]: E0321 08:51:56.592048 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d623e522-1d67-45ef-8ebd-a8a37e12f045" containerName="cloudkitty-api" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.592054 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d623e522-1d67-45ef-8ebd-a8a37e12f045" containerName="cloudkitty-api" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.592232 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="623a2b50-8a77-47df-800c-5ec7c237255c" containerName="cloudkitty-proc" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.592249 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="e75fe987-8f3c-4f76-9546-c6ab537af3ce" containerName="cloudkitty-storageinit" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.592258 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="817cb85b-e1e8-4ab7-93d9-2f7e78b8fc5a" containerName="registry-server" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.592271 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="d623e522-1d67-45ef-8ebd-a8a37e12f045" containerName="cloudkitty-api" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.592286 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="d623e522-1d67-45ef-8ebd-a8a37e12f045" containerName="cloudkitty-api-log" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.592296 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e77a97c-92ea-4bc5-bb15-b01d360220b6" containerName="dnsmasq-dns" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.593810 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.598658 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-public-svc" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.598831 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.598976 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-api-config-data" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.599052 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-internal-svc" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.599190 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.599411 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.598979 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-ff6v6" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.606735 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.626671 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-proc-0"] Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.628453 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.631346 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-proc-config-data" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.640198 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.700263 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.700434 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0413efc-468a-4065-9b3f-6355ec9d0e1e-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"e0413efc-468a-4065-9b3f-6355ec9d0e1e\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.700466 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0413efc-468a-4065-9b3f-6355ec9d0e1e-config-data\") pod \"cloudkitty-proc-0\" (UID: \"e0413efc-468a-4065-9b3f-6355ec9d0e1e\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.700510 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.700599 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.700632 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.700685 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0413efc-468a-4065-9b3f-6355ec9d0e1e-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"e0413efc-468a-4065-9b3f-6355ec9d0e1e\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.700715 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-config-data\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.700781 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-certs\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.700802 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-scripts\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.700874 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg8wv\" (UniqueName: \"kubernetes.io/projected/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-kube-api-access-xg8wv\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.700895 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/e0413efc-468a-4065-9b3f-6355ec9d0e1e-certs\") pod \"cloudkitty-proc-0\" (UID: \"e0413efc-468a-4065-9b3f-6355ec9d0e1e\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.700913 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldhxq\" (UniqueName: \"kubernetes.io/projected/e0413efc-468a-4065-9b3f-6355ec9d0e1e-kube-api-access-ldhxq\") pod \"cloudkitty-proc-0\" (UID: \"e0413efc-468a-4065-9b3f-6355ec9d0e1e\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.700948 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-logs\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.700965 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0413efc-468a-4065-9b3f-6355ec9d0e1e-scripts\") pod \"cloudkitty-proc-0\" (UID: \"e0413efc-468a-4065-9b3f-6355ec9d0e1e\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.804350 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0413efc-468a-4065-9b3f-6355ec9d0e1e-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"e0413efc-468a-4065-9b3f-6355ec9d0e1e\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.804416 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0413efc-468a-4065-9b3f-6355ec9d0e1e-config-data\") pod \"cloudkitty-proc-0\" (UID: \"e0413efc-468a-4065-9b3f-6355ec9d0e1e\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.804467 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.804490 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.804524 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.804581 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0413efc-468a-4065-9b3f-6355ec9d0e1e-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"e0413efc-468a-4065-9b3f-6355ec9d0e1e\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.804645 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-config-data\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.804681 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-certs\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.804703 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-scripts\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.804760 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg8wv\" (UniqueName: \"kubernetes.io/projected/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-kube-api-access-xg8wv\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.804784 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/e0413efc-468a-4065-9b3f-6355ec9d0e1e-certs\") pod \"cloudkitty-proc-0\" (UID: \"e0413efc-468a-4065-9b3f-6355ec9d0e1e\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.804804 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldhxq\" (UniqueName: \"kubernetes.io/projected/e0413efc-468a-4065-9b3f-6355ec9d0e1e-kube-api-access-ldhxq\") pod \"cloudkitty-proc-0\" (UID: \"e0413efc-468a-4065-9b3f-6355ec9d0e1e\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.804862 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-logs\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.804890 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0413efc-468a-4065-9b3f-6355ec9d0e1e-scripts\") pod \"cloudkitty-proc-0\" (UID: \"e0413efc-468a-4065-9b3f-6355ec9d0e1e\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.805052 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.806432 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-logs\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.811584 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.811689 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-certs\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.815431 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/e0413efc-468a-4065-9b3f-6355ec9d0e1e-certs\") pod \"cloudkitty-proc-0\" (UID: \"e0413efc-468a-4065-9b3f-6355ec9d0e1e\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.815464 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0413efc-468a-4065-9b3f-6355ec9d0e1e-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"e0413efc-468a-4065-9b3f-6355ec9d0e1e\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.815627 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.815986 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-config-data\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.816149 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0413efc-468a-4065-9b3f-6355ec9d0e1e-scripts\") pod \"cloudkitty-proc-0\" (UID: \"e0413efc-468a-4065-9b3f-6355ec9d0e1e\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.816720 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.817096 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-scripts\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.818102 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0413efc-468a-4065-9b3f-6355ec9d0e1e-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"e0413efc-468a-4065-9b3f-6355ec9d0e1e\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.823440 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg8wv\" (UniqueName: \"kubernetes.io/projected/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-kube-api-access-xg8wv\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.826894 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16a5bae2-5738-4f3c-a6ca-1e08aa8930a0-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0\") " pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.828337 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0413efc-468a-4065-9b3f-6355ec9d0e1e-config-data\") pod \"cloudkitty-proc-0\" (UID: \"e0413efc-468a-4065-9b3f-6355ec9d0e1e\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.828733 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldhxq\" (UniqueName: \"kubernetes.io/projected/e0413efc-468a-4065-9b3f-6355ec9d0e1e-kube-api-access-ldhxq\") pod \"cloudkitty-proc-0\" (UID: \"e0413efc-468a-4065-9b3f-6355ec9d0e1e\") " pod="openstack/cloudkitty-proc-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.920201 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Mar 21 08:51:56 crc kubenswrapper[4696]: I0321 08:51:56.949864 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Mar 21 08:51:57 crc kubenswrapper[4696]: W0321 08:51:57.483951 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16a5bae2_5738_4f3c_a6ca_1e08aa8930a0.slice/crio-0a92897dd6270ef56d2fb0ec6b1ed4681a120e18843b57c32ba6ddab2f16faf3 WatchSource:0}: Error finding container 0a92897dd6270ef56d2fb0ec6b1ed4681a120e18843b57c32ba6ddab2f16faf3: Status 404 returned error can't find the container with id 0a92897dd6270ef56d2fb0ec6b1ed4681a120e18843b57c32ba6ddab2f16faf3 Mar 21 08:51:57 crc kubenswrapper[4696]: I0321 08:51:57.495224 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Mar 21 08:51:57 crc kubenswrapper[4696]: I0321 08:51:57.528882 4696 scope.go:117] "RemoveContainer" containerID="3d200c0dfbaa3b07e65accea28350812647c92d7a181d86ec1e77b8666fdd45f" Mar 21 08:51:57 crc kubenswrapper[4696]: I0321 08:51:57.577007 4696 scope.go:117] "RemoveContainer" containerID="a85da9b90c784172d64500bb8c3d9906079ca9382be826d2a0cc79912eede670" Mar 21 08:51:57 crc kubenswrapper[4696]: I0321 08:51:57.577370 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Mar 21 08:51:58 crc kubenswrapper[4696]: I0321 08:51:58.475016 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0","Type":"ContainerStarted","Data":"3c960c192b91a7cbe598a37c0f35dd2272541ee9045bc10afd1fd8280f0b3b8f"} Mar 21 08:51:58 crc kubenswrapper[4696]: I0321 08:51:58.475657 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-api-0" Mar 21 08:51:58 crc kubenswrapper[4696]: I0321 08:51:58.475699 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0","Type":"ContainerStarted","Data":"1ac5b4693f33a37e949c0da79215633e20f3376f24b766f1983f82a57cfb5d9d"} Mar 21 08:51:58 crc kubenswrapper[4696]: I0321 08:51:58.475714 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"16a5bae2-5738-4f3c-a6ca-1e08aa8930a0","Type":"ContainerStarted","Data":"0a92897dd6270ef56d2fb0ec6b1ed4681a120e18843b57c32ba6ddab2f16faf3"} Mar 21 08:51:58 crc kubenswrapper[4696]: I0321 08:51:58.477370 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"e0413efc-468a-4065-9b3f-6355ec9d0e1e","Type":"ContainerStarted","Data":"63c955d6a77d0a5c630a3e34a48fa08a3d7d0648b646fd0df67d1cf3e6761ee7"} Mar 21 08:51:58 crc kubenswrapper[4696]: I0321 08:51:58.498325 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-api-0" podStartSLOduration=2.498307067 podStartE2EDuration="2.498307067s" podCreationTimestamp="2026-03-21 08:51:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:51:58.491329161 +0000 UTC m=+1452.612209874" watchObservedRunningTime="2026-03-21 08:51:58.498307067 +0000 UTC m=+1452.619187780" Mar 21 08:51:58 crc kubenswrapper[4696]: I0321 08:51:58.546402 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="623a2b50-8a77-47df-800c-5ec7c237255c" path="/var/lib/kubelet/pods/623a2b50-8a77-47df-800c-5ec7c237255c/volumes" Mar 21 08:51:59 crc kubenswrapper[4696]: I0321 08:51:59.488849 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"e0413efc-468a-4065-9b3f-6355ec9d0e1e","Type":"ContainerStarted","Data":"7652f82a6947d6f8121444c926579295c6a4c60bab9dab94ba688191fd2a4442"} Mar 21 08:51:59 crc kubenswrapper[4696]: I0321 08:51:59.515110 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-proc-0" podStartSLOduration=2.018335949 podStartE2EDuration="3.515069429s" podCreationTimestamp="2026-03-21 08:51:56 +0000 UTC" firstStartedPulling="2026-03-21 08:51:57.59174222 +0000 UTC m=+1451.712622933" lastFinishedPulling="2026-03-21 08:51:59.0884757 +0000 UTC m=+1453.209356413" observedRunningTime="2026-03-21 08:51:59.504917134 +0000 UTC m=+1453.625797867" watchObservedRunningTime="2026-03-21 08:51:59.515069429 +0000 UTC m=+1453.635950142" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.130483 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568052-n4r22"] Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.132256 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568052-n4r22" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.136779 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.137746 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.138561 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.140757 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568052-n4r22"] Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.185259 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzbzq\" (UniqueName: \"kubernetes.io/projected/0cdcde53-8a3a-449a-a525-fccb10e196aa-kube-api-access-jzbzq\") pod \"auto-csr-approver-29568052-n4r22\" (UID: \"0cdcde53-8a3a-449a-a525-fccb10e196aa\") " pod="openshift-infra/auto-csr-approver-29568052-n4r22" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.216503 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n"] Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.218050 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.220441 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.220585 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.220635 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.230051 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bqp9r" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.270695 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n"] Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.287780 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzbzq\" (UniqueName: \"kubernetes.io/projected/0cdcde53-8a3a-449a-a525-fccb10e196aa-kube-api-access-jzbzq\") pod \"auto-csr-approver-29568052-n4r22\" (UID: \"0cdcde53-8a3a-449a-a525-fccb10e196aa\") " pod="openshift-infra/auto-csr-approver-29568052-n4r22" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.287880 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/93d9988c-26af-48d1-b43e-7c897cad7235-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n\" (UID: \"93d9988c-26af-48d1-b43e-7c897cad7235\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.287961 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwppq\" (UniqueName: \"kubernetes.io/projected/93d9988c-26af-48d1-b43e-7c897cad7235-kube-api-access-pwppq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n\" (UID: \"93d9988c-26af-48d1-b43e-7c897cad7235\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.287995 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93d9988c-26af-48d1-b43e-7c897cad7235-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n\" (UID: \"93d9988c-26af-48d1-b43e-7c897cad7235\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.288362 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93d9988c-26af-48d1-b43e-7c897cad7235-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n\" (UID: \"93d9988c-26af-48d1-b43e-7c897cad7235\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.307779 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzbzq\" (UniqueName: \"kubernetes.io/projected/0cdcde53-8a3a-449a-a525-fccb10e196aa-kube-api-access-jzbzq\") pod \"auto-csr-approver-29568052-n4r22\" (UID: \"0cdcde53-8a3a-449a-a525-fccb10e196aa\") " pod="openshift-infra/auto-csr-approver-29568052-n4r22" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.390268 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93d9988c-26af-48d1-b43e-7c897cad7235-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n\" (UID: \"93d9988c-26af-48d1-b43e-7c897cad7235\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.390444 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/93d9988c-26af-48d1-b43e-7c897cad7235-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n\" (UID: \"93d9988c-26af-48d1-b43e-7c897cad7235\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.390549 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwppq\" (UniqueName: \"kubernetes.io/projected/93d9988c-26af-48d1-b43e-7c897cad7235-kube-api-access-pwppq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n\" (UID: \"93d9988c-26af-48d1-b43e-7c897cad7235\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.390598 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93d9988c-26af-48d1-b43e-7c897cad7235-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n\" (UID: \"93d9988c-26af-48d1-b43e-7c897cad7235\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.394564 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93d9988c-26af-48d1-b43e-7c897cad7235-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n\" (UID: \"93d9988c-26af-48d1-b43e-7c897cad7235\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.395421 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/93d9988c-26af-48d1-b43e-7c897cad7235-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n\" (UID: \"93d9988c-26af-48d1-b43e-7c897cad7235\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.395776 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93d9988c-26af-48d1-b43e-7c897cad7235-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n\" (UID: \"93d9988c-26af-48d1-b43e-7c897cad7235\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.407797 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwppq\" (UniqueName: \"kubernetes.io/projected/93d9988c-26af-48d1-b43e-7c897cad7235-kube-api-access-pwppq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n\" (UID: \"93d9988c-26af-48d1-b43e-7c897cad7235\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.457118 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568052-n4r22" Mar 21 08:52:00 crc kubenswrapper[4696]: I0321 08:52:00.533045 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n" Mar 21 08:52:01 crc kubenswrapper[4696]: I0321 08:52:01.251495 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568052-n4r22"] Mar 21 08:52:01 crc kubenswrapper[4696]: I0321 08:52:01.538314 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568052-n4r22" event={"ID":"0cdcde53-8a3a-449a-a525-fccb10e196aa","Type":"ContainerStarted","Data":"34ac904cc83eb5d60ae2ff8f1c21824f19f7083643a50619e2c1fbb05d319c36"} Mar 21 08:52:01 crc kubenswrapper[4696]: W0321 08:52:01.613453 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93d9988c_26af_48d1_b43e_7c897cad7235.slice/crio-24d6c8fbe5948ec1ff7331694f2c609dcf0f65a86796573f103e2212f441f935 WatchSource:0}: Error finding container 24d6c8fbe5948ec1ff7331694f2c609dcf0f65a86796573f103e2212f441f935: Status 404 returned error can't find the container with id 24d6c8fbe5948ec1ff7331694f2c609dcf0f65a86796573f103e2212f441f935 Mar 21 08:52:01 crc kubenswrapper[4696]: I0321 08:52:01.614317 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n"] Mar 21 08:52:02 crc kubenswrapper[4696]: I0321 08:52:02.552196 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n" event={"ID":"93d9988c-26af-48d1-b43e-7c897cad7235","Type":"ContainerStarted","Data":"24d6c8fbe5948ec1ff7331694f2c609dcf0f65a86796573f103e2212f441f935"} Mar 21 08:52:02 crc kubenswrapper[4696]: I0321 08:52:02.555751 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568052-n4r22" event={"ID":"0cdcde53-8a3a-449a-a525-fccb10e196aa","Type":"ContainerStarted","Data":"c90e0fb672ad228ae8aaa4a4b85436ed85fdd11eefb9c515d4a45dca0ae81b53"} Mar 21 08:52:02 crc kubenswrapper[4696]: I0321 08:52:02.579913 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29568052-n4r22" podStartSLOduration=1.747117308 podStartE2EDuration="2.579889553s" podCreationTimestamp="2026-03-21 08:52:00 +0000 UTC" firstStartedPulling="2026-03-21 08:52:01.277209883 +0000 UTC m=+1455.398090596" lastFinishedPulling="2026-03-21 08:52:02.109982128 +0000 UTC m=+1456.230862841" observedRunningTime="2026-03-21 08:52:02.568507024 +0000 UTC m=+1456.689387737" watchObservedRunningTime="2026-03-21 08:52:02.579889553 +0000 UTC m=+1456.700770266" Mar 21 08:52:03 crc kubenswrapper[4696]: I0321 08:52:03.574903 4696 generic.go:334] "Generic (PLEG): container finished" podID="0cdcde53-8a3a-449a-a525-fccb10e196aa" containerID="c90e0fb672ad228ae8aaa4a4b85436ed85fdd11eefb9c515d4a45dca0ae81b53" exitCode=0 Mar 21 08:52:03 crc kubenswrapper[4696]: I0321 08:52:03.574950 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568052-n4r22" event={"ID":"0cdcde53-8a3a-449a-a525-fccb10e196aa","Type":"ContainerDied","Data":"c90e0fb672ad228ae8aaa4a4b85436ed85fdd11eefb9c515d4a45dca0ae81b53"} Mar 21 08:52:05 crc kubenswrapper[4696]: I0321 08:52:05.373464 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568052-n4r22" Mar 21 08:52:05 crc kubenswrapper[4696]: I0321 08:52:05.455840 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzbzq\" (UniqueName: \"kubernetes.io/projected/0cdcde53-8a3a-449a-a525-fccb10e196aa-kube-api-access-jzbzq\") pod \"0cdcde53-8a3a-449a-a525-fccb10e196aa\" (UID: \"0cdcde53-8a3a-449a-a525-fccb10e196aa\") " Mar 21 08:52:05 crc kubenswrapper[4696]: I0321 08:52:05.461735 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cdcde53-8a3a-449a-a525-fccb10e196aa-kube-api-access-jzbzq" (OuterVolumeSpecName: "kube-api-access-jzbzq") pod "0cdcde53-8a3a-449a-a525-fccb10e196aa" (UID: "0cdcde53-8a3a-449a-a525-fccb10e196aa"). InnerVolumeSpecName "kube-api-access-jzbzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:52:05 crc kubenswrapper[4696]: I0321 08:52:05.558773 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzbzq\" (UniqueName: \"kubernetes.io/projected/0cdcde53-8a3a-449a-a525-fccb10e196aa-kube-api-access-jzbzq\") on node \"crc\" DevicePath \"\"" Mar 21 08:52:05 crc kubenswrapper[4696]: I0321 08:52:05.605842 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568052-n4r22" event={"ID":"0cdcde53-8a3a-449a-a525-fccb10e196aa","Type":"ContainerDied","Data":"34ac904cc83eb5d60ae2ff8f1c21824f19f7083643a50619e2c1fbb05d319c36"} Mar 21 08:52:05 crc kubenswrapper[4696]: I0321 08:52:05.605886 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34ac904cc83eb5d60ae2ff8f1c21824f19f7083643a50619e2c1fbb05d319c36" Mar 21 08:52:05 crc kubenswrapper[4696]: I0321 08:52:05.606000 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568052-n4r22" Mar 21 08:52:05 crc kubenswrapper[4696]: I0321 08:52:05.652861 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568046-b8pf4"] Mar 21 08:52:05 crc kubenswrapper[4696]: I0321 08:52:05.665283 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568046-b8pf4"] Mar 21 08:52:06 crc kubenswrapper[4696]: I0321 08:52:06.545940 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="579a841d-8410-4fa2-bc5f-14aafeaf17d6" path="/var/lib/kubelet/pods/579a841d-8410-4fa2-bc5f-14aafeaf17d6/volumes" Mar 21 08:52:06 crc kubenswrapper[4696]: I0321 08:52:06.617860 4696 generic.go:334] "Generic (PLEG): container finished" podID="56d6eab2-0e95-402a-afb7-2707ca41b144" containerID="80bc32955c6457e7b25214605028e3fda13557029d2ab6eb5265f8915dc62edc" exitCode=0 Mar 21 08:52:06 crc kubenswrapper[4696]: I0321 08:52:06.617959 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"56d6eab2-0e95-402a-afb7-2707ca41b144","Type":"ContainerDied","Data":"80bc32955c6457e7b25214605028e3fda13557029d2ab6eb5265f8915dc62edc"} Mar 21 08:52:06 crc kubenswrapper[4696]: I0321 08:52:06.619944 4696 generic.go:334] "Generic (PLEG): container finished" podID="0f40066c-3b43-42a5-9bb2-c8fb94a921db" containerID="591cc92872a6b54cb31429b395e1235f349c706bd3614cafb673fdfb11a4b1f1" exitCode=0 Mar 21 08:52:06 crc kubenswrapper[4696]: I0321 08:52:06.619975 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0f40066c-3b43-42a5-9bb2-c8fb94a921db","Type":"ContainerDied","Data":"591cc92872a6b54cb31429b395e1235f349c706bd3614cafb673fdfb11a4b1f1"} Mar 21 08:52:12 crc kubenswrapper[4696]: I0321 08:52:12.697683 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0f40066c-3b43-42a5-9bb2-c8fb94a921db","Type":"ContainerStarted","Data":"d08a99c94cdd18f562166c2d25b63abd2aa38faf6c7c58ad9cf96c723f680cc3"} Mar 21 08:52:12 crc kubenswrapper[4696]: I0321 08:52:12.699747 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:52:12 crc kubenswrapper[4696]: I0321 08:52:12.700145 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n" event={"ID":"93d9988c-26af-48d1-b43e-7c897cad7235","Type":"ContainerStarted","Data":"0bf62c58325f93726009cda36934c572e2a77d610e633e3d947227d7b255932c"} Mar 21 08:52:12 crc kubenswrapper[4696]: I0321 08:52:12.702328 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"56d6eab2-0e95-402a-afb7-2707ca41b144","Type":"ContainerStarted","Data":"0f8486249f6716f2613a27b6df2072cf308298d608ac4d8520e0f22c2a39a48d"} Mar 21 08:52:12 crc kubenswrapper[4696]: I0321 08:52:12.702594 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 21 08:52:12 crc kubenswrapper[4696]: I0321 08:52:12.730011 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=42.729970201 podStartE2EDuration="42.729970201s" podCreationTimestamp="2026-03-21 08:51:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:52:12.72708233 +0000 UTC m=+1466.847963043" watchObservedRunningTime="2026-03-21 08:52:12.729970201 +0000 UTC m=+1466.850850914" Mar 21 08:52:12 crc kubenswrapper[4696]: I0321 08:52:12.757113 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n" podStartSLOduration=2.4982297 podStartE2EDuration="12.757091612s" podCreationTimestamp="2026-03-21 08:52:00 +0000 UTC" firstStartedPulling="2026-03-21 08:52:01.615880223 +0000 UTC m=+1455.736760936" lastFinishedPulling="2026-03-21 08:52:11.874742135 +0000 UTC m=+1465.995622848" observedRunningTime="2026-03-21 08:52:12.756528137 +0000 UTC m=+1466.877408850" watchObservedRunningTime="2026-03-21 08:52:12.757091612 +0000 UTC m=+1466.877972315" Mar 21 08:52:12 crc kubenswrapper[4696]: I0321 08:52:12.787708 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=42.787684241 podStartE2EDuration="42.787684241s" podCreationTimestamp="2026-03-21 08:51:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 08:52:12.783638098 +0000 UTC m=+1466.904518811" watchObservedRunningTime="2026-03-21 08:52:12.787684241 +0000 UTC m=+1466.908564954" Mar 21 08:52:22 crc kubenswrapper[4696]: I0321 08:52:22.804187 4696 generic.go:334] "Generic (PLEG): container finished" podID="93d9988c-26af-48d1-b43e-7c897cad7235" containerID="0bf62c58325f93726009cda36934c572e2a77d610e633e3d947227d7b255932c" exitCode=0 Mar 21 08:52:22 crc kubenswrapper[4696]: I0321 08:52:22.804247 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n" event={"ID":"93d9988c-26af-48d1-b43e-7c897cad7235","Type":"ContainerDied","Data":"0bf62c58325f93726009cda36934c572e2a77d610e633e3d947227d7b255932c"} Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.347113 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n" Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.498750 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwppq\" (UniqueName: \"kubernetes.io/projected/93d9988c-26af-48d1-b43e-7c897cad7235-kube-api-access-pwppq\") pod \"93d9988c-26af-48d1-b43e-7c897cad7235\" (UID: \"93d9988c-26af-48d1-b43e-7c897cad7235\") " Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.498888 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93d9988c-26af-48d1-b43e-7c897cad7235-repo-setup-combined-ca-bundle\") pod \"93d9988c-26af-48d1-b43e-7c897cad7235\" (UID: \"93d9988c-26af-48d1-b43e-7c897cad7235\") " Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.498931 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/93d9988c-26af-48d1-b43e-7c897cad7235-ssh-key-openstack-edpm-ipam\") pod \"93d9988c-26af-48d1-b43e-7c897cad7235\" (UID: \"93d9988c-26af-48d1-b43e-7c897cad7235\") " Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.498951 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93d9988c-26af-48d1-b43e-7c897cad7235-inventory\") pod \"93d9988c-26af-48d1-b43e-7c897cad7235\" (UID: \"93d9988c-26af-48d1-b43e-7c897cad7235\") " Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.505359 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93d9988c-26af-48d1-b43e-7c897cad7235-kube-api-access-pwppq" (OuterVolumeSpecName: "kube-api-access-pwppq") pod "93d9988c-26af-48d1-b43e-7c897cad7235" (UID: "93d9988c-26af-48d1-b43e-7c897cad7235"). InnerVolumeSpecName "kube-api-access-pwppq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.505969 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93d9988c-26af-48d1-b43e-7c897cad7235-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "93d9988c-26af-48d1-b43e-7c897cad7235" (UID: "93d9988c-26af-48d1-b43e-7c897cad7235"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.529141 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93d9988c-26af-48d1-b43e-7c897cad7235-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "93d9988c-26af-48d1-b43e-7c897cad7235" (UID: "93d9988c-26af-48d1-b43e-7c897cad7235"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.538086 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93d9988c-26af-48d1-b43e-7c897cad7235-inventory" (OuterVolumeSpecName: "inventory") pod "93d9988c-26af-48d1-b43e-7c897cad7235" (UID: "93d9988c-26af-48d1-b43e-7c897cad7235"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.602749 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwppq\" (UniqueName: \"kubernetes.io/projected/93d9988c-26af-48d1-b43e-7c897cad7235-kube-api-access-pwppq\") on node \"crc\" DevicePath \"\"" Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.602842 4696 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93d9988c-26af-48d1-b43e-7c897cad7235-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.602870 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/93d9988c-26af-48d1-b43e-7c897cad7235-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.602903 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93d9988c-26af-48d1-b43e-7c897cad7235-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.853473 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n" Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.853473 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n" event={"ID":"93d9988c-26af-48d1-b43e-7c897cad7235","Type":"ContainerDied","Data":"24d6c8fbe5948ec1ff7331694f2c609dcf0f65a86796573f103e2212f441f935"} Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.856066 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24d6c8fbe5948ec1ff7331694f2c609dcf0f65a86796573f103e2212f441f935" Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.909404 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-6cdss"] Mar 21 08:52:24 crc kubenswrapper[4696]: E0321 08:52:24.909936 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93d9988c-26af-48d1-b43e-7c897cad7235" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.909954 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="93d9988c-26af-48d1-b43e-7c897cad7235" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 21 08:52:24 crc kubenswrapper[4696]: E0321 08:52:24.909978 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cdcde53-8a3a-449a-a525-fccb10e196aa" containerName="oc" Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.909984 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cdcde53-8a3a-449a-a525-fccb10e196aa" containerName="oc" Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.910184 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="93d9988c-26af-48d1-b43e-7c897cad7235" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.910209 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cdcde53-8a3a-449a-a525-fccb10e196aa" containerName="oc" Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.911025 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6cdss" Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.913086 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.913176 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.913458 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bqp9r" Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.914250 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 08:52:24 crc kubenswrapper[4696]: I0321 08:52:24.919288 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-6cdss"] Mar 21 08:52:25 crc kubenswrapper[4696]: I0321 08:52:25.012307 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/480a9890-311e-44b9-bfac-4c375e7cdf1e-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6cdss\" (UID: \"480a9890-311e-44b9-bfac-4c375e7cdf1e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6cdss" Mar 21 08:52:25 crc kubenswrapper[4696]: I0321 08:52:25.012396 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/480a9890-311e-44b9-bfac-4c375e7cdf1e-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6cdss\" (UID: \"480a9890-311e-44b9-bfac-4c375e7cdf1e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6cdss" Mar 21 08:52:25 crc kubenswrapper[4696]: I0321 08:52:25.012475 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9pm8\" (UniqueName: \"kubernetes.io/projected/480a9890-311e-44b9-bfac-4c375e7cdf1e-kube-api-access-l9pm8\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6cdss\" (UID: \"480a9890-311e-44b9-bfac-4c375e7cdf1e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6cdss" Mar 21 08:52:25 crc kubenswrapper[4696]: I0321 08:52:25.113951 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/480a9890-311e-44b9-bfac-4c375e7cdf1e-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6cdss\" (UID: \"480a9890-311e-44b9-bfac-4c375e7cdf1e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6cdss" Mar 21 08:52:25 crc kubenswrapper[4696]: I0321 08:52:25.114080 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/480a9890-311e-44b9-bfac-4c375e7cdf1e-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6cdss\" (UID: \"480a9890-311e-44b9-bfac-4c375e7cdf1e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6cdss" Mar 21 08:52:25 crc kubenswrapper[4696]: I0321 08:52:25.114211 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9pm8\" (UniqueName: \"kubernetes.io/projected/480a9890-311e-44b9-bfac-4c375e7cdf1e-kube-api-access-l9pm8\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6cdss\" (UID: \"480a9890-311e-44b9-bfac-4c375e7cdf1e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6cdss" Mar 21 08:52:25 crc kubenswrapper[4696]: I0321 08:52:25.118753 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/480a9890-311e-44b9-bfac-4c375e7cdf1e-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6cdss\" (UID: \"480a9890-311e-44b9-bfac-4c375e7cdf1e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6cdss" Mar 21 08:52:25 crc kubenswrapper[4696]: I0321 08:52:25.119680 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/480a9890-311e-44b9-bfac-4c375e7cdf1e-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6cdss\" (UID: \"480a9890-311e-44b9-bfac-4c375e7cdf1e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6cdss" Mar 21 08:52:25 crc kubenswrapper[4696]: I0321 08:52:25.131242 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9pm8\" (UniqueName: \"kubernetes.io/projected/480a9890-311e-44b9-bfac-4c375e7cdf1e-kube-api-access-l9pm8\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6cdss\" (UID: \"480a9890-311e-44b9-bfac-4c375e7cdf1e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6cdss" Mar 21 08:52:25 crc kubenswrapper[4696]: I0321 08:52:25.228271 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6cdss" Mar 21 08:52:25 crc kubenswrapper[4696]: I0321 08:52:25.917926 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-6cdss"] Mar 21 08:52:25 crc kubenswrapper[4696]: W0321 08:52:25.920059 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod480a9890_311e_44b9_bfac_4c375e7cdf1e.slice/crio-6bb5cd25f32ccf9e466b20a8f9571f0fdf7c2ad682e9e3dcc2b5d7515b715130 WatchSource:0}: Error finding container 6bb5cd25f32ccf9e466b20a8f9571f0fdf7c2ad682e9e3dcc2b5d7515b715130: Status 404 returned error can't find the container with id 6bb5cd25f32ccf9e466b20a8f9571f0fdf7c2ad682e9e3dcc2b5d7515b715130 Mar 21 08:52:26 crc kubenswrapper[4696]: I0321 08:52:26.885886 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6cdss" event={"ID":"480a9890-311e-44b9-bfac-4c375e7cdf1e","Type":"ContainerStarted","Data":"875f32b6897842a5e13e8328426906effa396dc4e09891e1c72d30fec89e9657"} Mar 21 08:52:26 crc kubenswrapper[4696]: I0321 08:52:26.886557 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6cdss" event={"ID":"480a9890-311e-44b9-bfac-4c375e7cdf1e","Type":"ContainerStarted","Data":"6bb5cd25f32ccf9e466b20a8f9571f0fdf7c2ad682e9e3dcc2b5d7515b715130"} Mar 21 08:52:26 crc kubenswrapper[4696]: I0321 08:52:26.911857 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6cdss" podStartSLOduration=2.518159319 podStartE2EDuration="2.911833134s" podCreationTimestamp="2026-03-21 08:52:24 +0000 UTC" firstStartedPulling="2026-03-21 08:52:25.924158718 +0000 UTC m=+1480.045039431" lastFinishedPulling="2026-03-21 08:52:26.317832533 +0000 UTC m=+1480.438713246" observedRunningTime="2026-03-21 08:52:26.902311496 +0000 UTC m=+1481.023192209" watchObservedRunningTime="2026-03-21 08:52:26.911833134 +0000 UTC m=+1481.032713847" Mar 21 08:52:29 crc kubenswrapper[4696]: I0321 08:52:29.922622 4696 generic.go:334] "Generic (PLEG): container finished" podID="480a9890-311e-44b9-bfac-4c375e7cdf1e" containerID="875f32b6897842a5e13e8328426906effa396dc4e09891e1c72d30fec89e9657" exitCode=0 Mar 21 08:52:29 crc kubenswrapper[4696]: I0321 08:52:29.922943 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6cdss" event={"ID":"480a9890-311e-44b9-bfac-4c375e7cdf1e","Type":"ContainerDied","Data":"875f32b6897842a5e13e8328426906effa396dc4e09891e1c72d30fec89e9657"} Mar 21 08:52:30 crc kubenswrapper[4696]: I0321 08:52:30.622147 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 21 08:52:30 crc kubenswrapper[4696]: I0321 08:52:30.643006 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 21 08:52:31 crc kubenswrapper[4696]: I0321 08:52:31.518710 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6cdss" Mar 21 08:52:31 crc kubenswrapper[4696]: I0321 08:52:31.648338 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/480a9890-311e-44b9-bfac-4c375e7cdf1e-ssh-key-openstack-edpm-ipam\") pod \"480a9890-311e-44b9-bfac-4c375e7cdf1e\" (UID: \"480a9890-311e-44b9-bfac-4c375e7cdf1e\") " Mar 21 08:52:31 crc kubenswrapper[4696]: I0321 08:52:31.648490 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/480a9890-311e-44b9-bfac-4c375e7cdf1e-inventory\") pod \"480a9890-311e-44b9-bfac-4c375e7cdf1e\" (UID: \"480a9890-311e-44b9-bfac-4c375e7cdf1e\") " Mar 21 08:52:31 crc kubenswrapper[4696]: I0321 08:52:31.648615 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9pm8\" (UniqueName: \"kubernetes.io/projected/480a9890-311e-44b9-bfac-4c375e7cdf1e-kube-api-access-l9pm8\") pod \"480a9890-311e-44b9-bfac-4c375e7cdf1e\" (UID: \"480a9890-311e-44b9-bfac-4c375e7cdf1e\") " Mar 21 08:52:31 crc kubenswrapper[4696]: I0321 08:52:31.656164 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/480a9890-311e-44b9-bfac-4c375e7cdf1e-kube-api-access-l9pm8" (OuterVolumeSpecName: "kube-api-access-l9pm8") pod "480a9890-311e-44b9-bfac-4c375e7cdf1e" (UID: "480a9890-311e-44b9-bfac-4c375e7cdf1e"). InnerVolumeSpecName "kube-api-access-l9pm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:52:31 crc kubenswrapper[4696]: I0321 08:52:31.678366 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/480a9890-311e-44b9-bfac-4c375e7cdf1e-inventory" (OuterVolumeSpecName: "inventory") pod "480a9890-311e-44b9-bfac-4c375e7cdf1e" (UID: "480a9890-311e-44b9-bfac-4c375e7cdf1e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:52:31 crc kubenswrapper[4696]: I0321 08:52:31.684929 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/480a9890-311e-44b9-bfac-4c375e7cdf1e-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "480a9890-311e-44b9-bfac-4c375e7cdf1e" (UID: "480a9890-311e-44b9-bfac-4c375e7cdf1e"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:52:31 crc kubenswrapper[4696]: I0321 08:52:31.751686 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/480a9890-311e-44b9-bfac-4c375e7cdf1e-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 08:52:31 crc kubenswrapper[4696]: I0321 08:52:31.751721 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9pm8\" (UniqueName: \"kubernetes.io/projected/480a9890-311e-44b9-bfac-4c375e7cdf1e-kube-api-access-l9pm8\") on node \"crc\" DevicePath \"\"" Mar 21 08:52:31 crc kubenswrapper[4696]: I0321 08:52:31.751731 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/480a9890-311e-44b9-bfac-4c375e7cdf1e-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 08:52:31 crc kubenswrapper[4696]: I0321 08:52:31.957319 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6cdss" event={"ID":"480a9890-311e-44b9-bfac-4c375e7cdf1e","Type":"ContainerDied","Data":"6bb5cd25f32ccf9e466b20a8f9571f0fdf7c2ad682e9e3dcc2b5d7515b715130"} Mar 21 08:52:31 crc kubenswrapper[4696]: I0321 08:52:31.957358 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6bb5cd25f32ccf9e466b20a8f9571f0fdf7c2ad682e9e3dcc2b5d7515b715130" Mar 21 08:52:31 crc kubenswrapper[4696]: I0321 08:52:31.957417 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6cdss" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.030093 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq"] Mar 21 08:52:32 crc kubenswrapper[4696]: E0321 08:52:32.031340 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480a9890-311e-44b9-bfac-4c375e7cdf1e" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.031367 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="480a9890-311e-44b9-bfac-4c375e7cdf1e" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.031662 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="480a9890-311e-44b9-bfac-4c375e7cdf1e" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.033083 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.035630 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.035914 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.036196 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bqp9r" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.036385 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.045430 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq"] Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.160396 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/875486e6-3369-4601-bc53-00c2aaf85707-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq\" (UID: \"875486e6-3369-4601-bc53-00c2aaf85707\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.160641 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/875486e6-3369-4601-bc53-00c2aaf85707-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq\" (UID: \"875486e6-3369-4601-bc53-00c2aaf85707\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.160694 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/875486e6-3369-4601-bc53-00c2aaf85707-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq\" (UID: \"875486e6-3369-4601-bc53-00c2aaf85707\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.160730 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6pl2\" (UniqueName: \"kubernetes.io/projected/875486e6-3369-4601-bc53-00c2aaf85707-kube-api-access-s6pl2\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq\" (UID: \"875486e6-3369-4601-bc53-00c2aaf85707\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.264071 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/875486e6-3369-4601-bc53-00c2aaf85707-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq\" (UID: \"875486e6-3369-4601-bc53-00c2aaf85707\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.264185 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/875486e6-3369-4601-bc53-00c2aaf85707-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq\" (UID: \"875486e6-3369-4601-bc53-00c2aaf85707\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.264232 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6pl2\" (UniqueName: \"kubernetes.io/projected/875486e6-3369-4601-bc53-00c2aaf85707-kube-api-access-s6pl2\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq\" (UID: \"875486e6-3369-4601-bc53-00c2aaf85707\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.264273 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/875486e6-3369-4601-bc53-00c2aaf85707-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq\" (UID: \"875486e6-3369-4601-bc53-00c2aaf85707\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.268896 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/875486e6-3369-4601-bc53-00c2aaf85707-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq\" (UID: \"875486e6-3369-4601-bc53-00c2aaf85707\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.269128 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/875486e6-3369-4601-bc53-00c2aaf85707-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq\" (UID: \"875486e6-3369-4601-bc53-00c2aaf85707\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.278994 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/875486e6-3369-4601-bc53-00c2aaf85707-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq\" (UID: \"875486e6-3369-4601-bc53-00c2aaf85707\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.283848 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6pl2\" (UniqueName: \"kubernetes.io/projected/875486e6-3369-4601-bc53-00c2aaf85707-kube-api-access-s6pl2\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq\" (UID: \"875486e6-3369-4601-bc53-00c2aaf85707\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.385313 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq" Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.920111 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq"] Mar 21 08:52:32 crc kubenswrapper[4696]: I0321 08:52:32.968759 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq" event={"ID":"875486e6-3369-4601-bc53-00c2aaf85707","Type":"ContainerStarted","Data":"811538461d3743dc98dea22cdf84f6a5ea72e332c1f86863bf87582bd85c4617"} Mar 21 08:52:33 crc kubenswrapper[4696]: I0321 08:52:33.917590 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-api-0" Mar 21 08:52:35 crc kubenswrapper[4696]: I0321 08:52:35.997994 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq" event={"ID":"875486e6-3369-4601-bc53-00c2aaf85707","Type":"ContainerStarted","Data":"1d2e2ff8b196017f4c18a9283358e136c6ea0eab6e1251edb32e8f96e2e90cf3"} Mar 21 08:52:36 crc kubenswrapper[4696]: I0321 08:52:36.039353 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq" podStartSLOduration=2.213961107 podStartE2EDuration="4.039336806s" podCreationTimestamp="2026-03-21 08:52:32 +0000 UTC" firstStartedPulling="2026-03-21 08:52:32.921898004 +0000 UTC m=+1487.042778717" lastFinishedPulling="2026-03-21 08:52:34.747273703 +0000 UTC m=+1488.868154416" observedRunningTime="2026-03-21 08:52:36.017566814 +0000 UTC m=+1490.138447547" watchObservedRunningTime="2026-03-21 08:52:36.039336806 +0000 UTC m=+1490.160217519" Mar 21 08:52:57 crc kubenswrapper[4696]: I0321 08:52:57.931421 4696 scope.go:117] "RemoveContainer" containerID="ebe25e48963bc833c63e1295302833398482ec0ae4b9f85c9c5e746501f31a64" Mar 21 08:52:57 crc kubenswrapper[4696]: I0321 08:52:57.959105 4696 scope.go:117] "RemoveContainer" containerID="dbc48018407d29c23bb500596ff5dfd1c996e0eee22ffe4e1a86f17995f71a41" Mar 21 08:52:58 crc kubenswrapper[4696]: I0321 08:52:58.018783 4696 scope.go:117] "RemoveContainer" containerID="00f2cf65b6102c22b225f29f98f3f9ac8a2f45e936cac062139f4b8e3a6a2594" Mar 21 08:52:58 crc kubenswrapper[4696]: I0321 08:52:58.075260 4696 scope.go:117] "RemoveContainer" containerID="90a6920f56317148709fbc10994ed7dfcbba2b030c0f613beef00e4bbe2fa4b2" Mar 21 08:53:10 crc kubenswrapper[4696]: I0321 08:53:10.545399 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k2nfq"] Mar 21 08:53:10 crc kubenswrapper[4696]: I0321 08:53:10.547847 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2nfq" Mar 21 08:53:10 crc kubenswrapper[4696]: I0321 08:53:10.552109 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2nfq"] Mar 21 08:53:10 crc kubenswrapper[4696]: I0321 08:53:10.659153 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59gh8\" (UniqueName: \"kubernetes.io/projected/82dc20dc-9107-47bd-8079-757f8b1161dd-kube-api-access-59gh8\") pod \"redhat-marketplace-k2nfq\" (UID: \"82dc20dc-9107-47bd-8079-757f8b1161dd\") " pod="openshift-marketplace/redhat-marketplace-k2nfq" Mar 21 08:53:10 crc kubenswrapper[4696]: I0321 08:53:10.659326 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82dc20dc-9107-47bd-8079-757f8b1161dd-catalog-content\") pod \"redhat-marketplace-k2nfq\" (UID: \"82dc20dc-9107-47bd-8079-757f8b1161dd\") " pod="openshift-marketplace/redhat-marketplace-k2nfq" Mar 21 08:53:10 crc kubenswrapper[4696]: I0321 08:53:10.659732 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82dc20dc-9107-47bd-8079-757f8b1161dd-utilities\") pod \"redhat-marketplace-k2nfq\" (UID: \"82dc20dc-9107-47bd-8079-757f8b1161dd\") " pod="openshift-marketplace/redhat-marketplace-k2nfq" Mar 21 08:53:10 crc kubenswrapper[4696]: I0321 08:53:10.761993 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82dc20dc-9107-47bd-8079-757f8b1161dd-catalog-content\") pod \"redhat-marketplace-k2nfq\" (UID: \"82dc20dc-9107-47bd-8079-757f8b1161dd\") " pod="openshift-marketplace/redhat-marketplace-k2nfq" Mar 21 08:53:10 crc kubenswrapper[4696]: I0321 08:53:10.762133 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82dc20dc-9107-47bd-8079-757f8b1161dd-utilities\") pod \"redhat-marketplace-k2nfq\" (UID: \"82dc20dc-9107-47bd-8079-757f8b1161dd\") " pod="openshift-marketplace/redhat-marketplace-k2nfq" Mar 21 08:53:10 crc kubenswrapper[4696]: I0321 08:53:10.762258 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59gh8\" (UniqueName: \"kubernetes.io/projected/82dc20dc-9107-47bd-8079-757f8b1161dd-kube-api-access-59gh8\") pod \"redhat-marketplace-k2nfq\" (UID: \"82dc20dc-9107-47bd-8079-757f8b1161dd\") " pod="openshift-marketplace/redhat-marketplace-k2nfq" Mar 21 08:53:10 crc kubenswrapper[4696]: I0321 08:53:10.762582 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82dc20dc-9107-47bd-8079-757f8b1161dd-catalog-content\") pod \"redhat-marketplace-k2nfq\" (UID: \"82dc20dc-9107-47bd-8079-757f8b1161dd\") " pod="openshift-marketplace/redhat-marketplace-k2nfq" Mar 21 08:53:10 crc kubenswrapper[4696]: I0321 08:53:10.762675 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82dc20dc-9107-47bd-8079-757f8b1161dd-utilities\") pod \"redhat-marketplace-k2nfq\" (UID: \"82dc20dc-9107-47bd-8079-757f8b1161dd\") " pod="openshift-marketplace/redhat-marketplace-k2nfq" Mar 21 08:53:10 crc kubenswrapper[4696]: I0321 08:53:10.782124 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59gh8\" (UniqueName: \"kubernetes.io/projected/82dc20dc-9107-47bd-8079-757f8b1161dd-kube-api-access-59gh8\") pod \"redhat-marketplace-k2nfq\" (UID: \"82dc20dc-9107-47bd-8079-757f8b1161dd\") " pod="openshift-marketplace/redhat-marketplace-k2nfq" Mar 21 08:53:10 crc kubenswrapper[4696]: I0321 08:53:10.869363 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2nfq" Mar 21 08:53:11 crc kubenswrapper[4696]: I0321 08:53:11.555743 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2nfq"] Mar 21 08:53:11 crc kubenswrapper[4696]: I0321 08:53:11.720168 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2nfq" event={"ID":"82dc20dc-9107-47bd-8079-757f8b1161dd","Type":"ContainerStarted","Data":"47ec03aa9c369cecd1cbfad49e057c36c79cd41dbdfd916b19c645ad1f44edc5"} Mar 21 08:53:12 crc kubenswrapper[4696]: I0321 08:53:12.733390 4696 generic.go:334] "Generic (PLEG): container finished" podID="82dc20dc-9107-47bd-8079-757f8b1161dd" containerID="ad2e8c5c14fac8910956eada61c1b63ec1b01816a3a0ebbf9e65bb6464b7b818" exitCode=0 Mar 21 08:53:12 crc kubenswrapper[4696]: I0321 08:53:12.733455 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2nfq" event={"ID":"82dc20dc-9107-47bd-8079-757f8b1161dd","Type":"ContainerDied","Data":"ad2e8c5c14fac8910956eada61c1b63ec1b01816a3a0ebbf9e65bb6464b7b818"} Mar 21 08:53:13 crc kubenswrapper[4696]: I0321 08:53:13.745381 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2nfq" event={"ID":"82dc20dc-9107-47bd-8079-757f8b1161dd","Type":"ContainerStarted","Data":"d7b5850909749b26d20ed183281b9f8abbc4b470bdecb833da7c3bb127b98adb"} Mar 21 08:53:14 crc kubenswrapper[4696]: I0321 08:53:14.757392 4696 generic.go:334] "Generic (PLEG): container finished" podID="82dc20dc-9107-47bd-8079-757f8b1161dd" containerID="d7b5850909749b26d20ed183281b9f8abbc4b470bdecb833da7c3bb127b98adb" exitCode=0 Mar 21 08:53:14 crc kubenswrapper[4696]: I0321 08:53:14.757508 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2nfq" event={"ID":"82dc20dc-9107-47bd-8079-757f8b1161dd","Type":"ContainerDied","Data":"d7b5850909749b26d20ed183281b9f8abbc4b470bdecb833da7c3bb127b98adb"} Mar 21 08:53:15 crc kubenswrapper[4696]: I0321 08:53:15.768875 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2nfq" event={"ID":"82dc20dc-9107-47bd-8079-757f8b1161dd","Type":"ContainerStarted","Data":"1732fc6db6d5d896bf7317ffecdfe4b3eb6d164ad2cc9e0360bb6eb1c0aa8ef4"} Mar 21 08:53:15 crc kubenswrapper[4696]: I0321 08:53:15.798020 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k2nfq" podStartSLOduration=3.3661606490000002 podStartE2EDuration="5.798002508s" podCreationTimestamp="2026-03-21 08:53:10 +0000 UTC" firstStartedPulling="2026-03-21 08:53:12.735723385 +0000 UTC m=+1526.856604098" lastFinishedPulling="2026-03-21 08:53:15.167565244 +0000 UTC m=+1529.288445957" observedRunningTime="2026-03-21 08:53:15.785195737 +0000 UTC m=+1529.906076450" watchObservedRunningTime="2026-03-21 08:53:15.798002508 +0000 UTC m=+1529.918883221" Mar 21 08:53:20 crc kubenswrapper[4696]: I0321 08:53:20.870080 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k2nfq" Mar 21 08:53:20 crc kubenswrapper[4696]: I0321 08:53:20.870479 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k2nfq" Mar 21 08:53:20 crc kubenswrapper[4696]: I0321 08:53:20.950229 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k2nfq" Mar 21 08:53:21 crc kubenswrapper[4696]: I0321 08:53:21.887302 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k2nfq" Mar 21 08:53:21 crc kubenswrapper[4696]: I0321 08:53:21.945175 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2nfq"] Mar 21 08:53:23 crc kubenswrapper[4696]: I0321 08:53:23.851019 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-k2nfq" podUID="82dc20dc-9107-47bd-8079-757f8b1161dd" containerName="registry-server" containerID="cri-o://1732fc6db6d5d896bf7317ffecdfe4b3eb6d164ad2cc9e0360bb6eb1c0aa8ef4" gracePeriod=2 Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.700544 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2nfq" Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.768520 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82dc20dc-9107-47bd-8079-757f8b1161dd-utilities\") pod \"82dc20dc-9107-47bd-8079-757f8b1161dd\" (UID: \"82dc20dc-9107-47bd-8079-757f8b1161dd\") " Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.768565 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82dc20dc-9107-47bd-8079-757f8b1161dd-catalog-content\") pod \"82dc20dc-9107-47bd-8079-757f8b1161dd\" (UID: \"82dc20dc-9107-47bd-8079-757f8b1161dd\") " Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.768634 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59gh8\" (UniqueName: \"kubernetes.io/projected/82dc20dc-9107-47bd-8079-757f8b1161dd-kube-api-access-59gh8\") pod \"82dc20dc-9107-47bd-8079-757f8b1161dd\" (UID: \"82dc20dc-9107-47bd-8079-757f8b1161dd\") " Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.769371 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82dc20dc-9107-47bd-8079-757f8b1161dd-utilities" (OuterVolumeSpecName: "utilities") pod "82dc20dc-9107-47bd-8079-757f8b1161dd" (UID: "82dc20dc-9107-47bd-8079-757f8b1161dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.779560 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82dc20dc-9107-47bd-8079-757f8b1161dd-kube-api-access-59gh8" (OuterVolumeSpecName: "kube-api-access-59gh8") pod "82dc20dc-9107-47bd-8079-757f8b1161dd" (UID: "82dc20dc-9107-47bd-8079-757f8b1161dd"). InnerVolumeSpecName "kube-api-access-59gh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.795978 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82dc20dc-9107-47bd-8079-757f8b1161dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "82dc20dc-9107-47bd-8079-757f8b1161dd" (UID: "82dc20dc-9107-47bd-8079-757f8b1161dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.870544 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82dc20dc-9107-47bd-8079-757f8b1161dd-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.870569 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82dc20dc-9107-47bd-8079-757f8b1161dd-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.870580 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59gh8\" (UniqueName: \"kubernetes.io/projected/82dc20dc-9107-47bd-8079-757f8b1161dd-kube-api-access-59gh8\") on node \"crc\" DevicePath \"\"" Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.873969 4696 generic.go:334] "Generic (PLEG): container finished" podID="82dc20dc-9107-47bd-8079-757f8b1161dd" containerID="1732fc6db6d5d896bf7317ffecdfe4b3eb6d164ad2cc9e0360bb6eb1c0aa8ef4" exitCode=0 Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.874003 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2nfq" event={"ID":"82dc20dc-9107-47bd-8079-757f8b1161dd","Type":"ContainerDied","Data":"1732fc6db6d5d896bf7317ffecdfe4b3eb6d164ad2cc9e0360bb6eb1c0aa8ef4"} Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.874028 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2nfq" event={"ID":"82dc20dc-9107-47bd-8079-757f8b1161dd","Type":"ContainerDied","Data":"47ec03aa9c369cecd1cbfad49e057c36c79cd41dbdfd916b19c645ad1f44edc5"} Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.874044 4696 scope.go:117] "RemoveContainer" containerID="1732fc6db6d5d896bf7317ffecdfe4b3eb6d164ad2cc9e0360bb6eb1c0aa8ef4" Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.874153 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2nfq" Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.904746 4696 scope.go:117] "RemoveContainer" containerID="d7b5850909749b26d20ed183281b9f8abbc4b470bdecb833da7c3bb127b98adb" Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.913190 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2nfq"] Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.929892 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2nfq"] Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.963223 4696 scope.go:117] "RemoveContainer" containerID="ad2e8c5c14fac8910956eada61c1b63ec1b01816a3a0ebbf9e65bb6464b7b818" Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.990813 4696 scope.go:117] "RemoveContainer" containerID="1732fc6db6d5d896bf7317ffecdfe4b3eb6d164ad2cc9e0360bb6eb1c0aa8ef4" Mar 21 08:53:24 crc kubenswrapper[4696]: E0321 08:53:24.991302 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1732fc6db6d5d896bf7317ffecdfe4b3eb6d164ad2cc9e0360bb6eb1c0aa8ef4\": container with ID starting with 1732fc6db6d5d896bf7317ffecdfe4b3eb6d164ad2cc9e0360bb6eb1c0aa8ef4 not found: ID does not exist" containerID="1732fc6db6d5d896bf7317ffecdfe4b3eb6d164ad2cc9e0360bb6eb1c0aa8ef4" Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.991355 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1732fc6db6d5d896bf7317ffecdfe4b3eb6d164ad2cc9e0360bb6eb1c0aa8ef4"} err="failed to get container status \"1732fc6db6d5d896bf7317ffecdfe4b3eb6d164ad2cc9e0360bb6eb1c0aa8ef4\": rpc error: code = NotFound desc = could not find container \"1732fc6db6d5d896bf7317ffecdfe4b3eb6d164ad2cc9e0360bb6eb1c0aa8ef4\": container with ID starting with 1732fc6db6d5d896bf7317ffecdfe4b3eb6d164ad2cc9e0360bb6eb1c0aa8ef4 not found: ID does not exist" Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.991383 4696 scope.go:117] "RemoveContainer" containerID="d7b5850909749b26d20ed183281b9f8abbc4b470bdecb833da7c3bb127b98adb" Mar 21 08:53:24 crc kubenswrapper[4696]: E0321 08:53:24.991878 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7b5850909749b26d20ed183281b9f8abbc4b470bdecb833da7c3bb127b98adb\": container with ID starting with d7b5850909749b26d20ed183281b9f8abbc4b470bdecb833da7c3bb127b98adb not found: ID does not exist" containerID="d7b5850909749b26d20ed183281b9f8abbc4b470bdecb833da7c3bb127b98adb" Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.991909 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7b5850909749b26d20ed183281b9f8abbc4b470bdecb833da7c3bb127b98adb"} err="failed to get container status \"d7b5850909749b26d20ed183281b9f8abbc4b470bdecb833da7c3bb127b98adb\": rpc error: code = NotFound desc = could not find container \"d7b5850909749b26d20ed183281b9f8abbc4b470bdecb833da7c3bb127b98adb\": container with ID starting with d7b5850909749b26d20ed183281b9f8abbc4b470bdecb833da7c3bb127b98adb not found: ID does not exist" Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.991932 4696 scope.go:117] "RemoveContainer" containerID="ad2e8c5c14fac8910956eada61c1b63ec1b01816a3a0ebbf9e65bb6464b7b818" Mar 21 08:53:24 crc kubenswrapper[4696]: E0321 08:53:24.992186 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad2e8c5c14fac8910956eada61c1b63ec1b01816a3a0ebbf9e65bb6464b7b818\": container with ID starting with ad2e8c5c14fac8910956eada61c1b63ec1b01816a3a0ebbf9e65bb6464b7b818 not found: ID does not exist" containerID="ad2e8c5c14fac8910956eada61c1b63ec1b01816a3a0ebbf9e65bb6464b7b818" Mar 21 08:53:24 crc kubenswrapper[4696]: I0321 08:53:24.992205 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad2e8c5c14fac8910956eada61c1b63ec1b01816a3a0ebbf9e65bb6464b7b818"} err="failed to get container status \"ad2e8c5c14fac8910956eada61c1b63ec1b01816a3a0ebbf9e65bb6464b7b818\": rpc error: code = NotFound desc = could not find container \"ad2e8c5c14fac8910956eada61c1b63ec1b01816a3a0ebbf9e65bb6464b7b818\": container with ID starting with ad2e8c5c14fac8910956eada61c1b63ec1b01816a3a0ebbf9e65bb6464b7b818 not found: ID does not exist" Mar 21 08:53:26 crc kubenswrapper[4696]: I0321 08:53:26.547349 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82dc20dc-9107-47bd-8079-757f8b1161dd" path="/var/lib/kubelet/pods/82dc20dc-9107-47bd-8079-757f8b1161dd/volumes" Mar 21 08:53:50 crc kubenswrapper[4696]: I0321 08:53:50.814859 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m8kv6"] Mar 21 08:53:50 crc kubenswrapper[4696]: E0321 08:53:50.815987 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82dc20dc-9107-47bd-8079-757f8b1161dd" containerName="extract-utilities" Mar 21 08:53:50 crc kubenswrapper[4696]: I0321 08:53:50.816008 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="82dc20dc-9107-47bd-8079-757f8b1161dd" containerName="extract-utilities" Mar 21 08:53:50 crc kubenswrapper[4696]: E0321 08:53:50.816056 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82dc20dc-9107-47bd-8079-757f8b1161dd" containerName="registry-server" Mar 21 08:53:50 crc kubenswrapper[4696]: I0321 08:53:50.816067 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="82dc20dc-9107-47bd-8079-757f8b1161dd" containerName="registry-server" Mar 21 08:53:50 crc kubenswrapper[4696]: E0321 08:53:50.816085 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82dc20dc-9107-47bd-8079-757f8b1161dd" containerName="extract-content" Mar 21 08:53:50 crc kubenswrapper[4696]: I0321 08:53:50.816092 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="82dc20dc-9107-47bd-8079-757f8b1161dd" containerName="extract-content" Mar 21 08:53:50 crc kubenswrapper[4696]: I0321 08:53:50.816362 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="82dc20dc-9107-47bd-8079-757f8b1161dd" containerName="registry-server" Mar 21 08:53:50 crc kubenswrapper[4696]: I0321 08:53:50.818407 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m8kv6" Mar 21 08:53:50 crc kubenswrapper[4696]: I0321 08:53:50.846161 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m8kv6"] Mar 21 08:53:50 crc kubenswrapper[4696]: I0321 08:53:50.890044 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a098079d-37fa-4f22-a9aa-ee57589c52ab-catalog-content\") pod \"certified-operators-m8kv6\" (UID: \"a098079d-37fa-4f22-a9aa-ee57589c52ab\") " pod="openshift-marketplace/certified-operators-m8kv6" Mar 21 08:53:50 crc kubenswrapper[4696]: I0321 08:53:50.890249 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pz8v\" (UniqueName: \"kubernetes.io/projected/a098079d-37fa-4f22-a9aa-ee57589c52ab-kube-api-access-8pz8v\") pod \"certified-operators-m8kv6\" (UID: \"a098079d-37fa-4f22-a9aa-ee57589c52ab\") " pod="openshift-marketplace/certified-operators-m8kv6" Mar 21 08:53:50 crc kubenswrapper[4696]: I0321 08:53:50.890709 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a098079d-37fa-4f22-a9aa-ee57589c52ab-utilities\") pod \"certified-operators-m8kv6\" (UID: \"a098079d-37fa-4f22-a9aa-ee57589c52ab\") " pod="openshift-marketplace/certified-operators-m8kv6" Mar 21 08:53:50 crc kubenswrapper[4696]: I0321 08:53:50.992706 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a098079d-37fa-4f22-a9aa-ee57589c52ab-utilities\") pod \"certified-operators-m8kv6\" (UID: \"a098079d-37fa-4f22-a9aa-ee57589c52ab\") " pod="openshift-marketplace/certified-operators-m8kv6" Mar 21 08:53:50 crc kubenswrapper[4696]: I0321 08:53:50.993154 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a098079d-37fa-4f22-a9aa-ee57589c52ab-catalog-content\") pod \"certified-operators-m8kv6\" (UID: \"a098079d-37fa-4f22-a9aa-ee57589c52ab\") " pod="openshift-marketplace/certified-operators-m8kv6" Mar 21 08:53:50 crc kubenswrapper[4696]: I0321 08:53:50.993214 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pz8v\" (UniqueName: \"kubernetes.io/projected/a098079d-37fa-4f22-a9aa-ee57589c52ab-kube-api-access-8pz8v\") pod \"certified-operators-m8kv6\" (UID: \"a098079d-37fa-4f22-a9aa-ee57589c52ab\") " pod="openshift-marketplace/certified-operators-m8kv6" Mar 21 08:53:50 crc kubenswrapper[4696]: I0321 08:53:50.993279 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a098079d-37fa-4f22-a9aa-ee57589c52ab-utilities\") pod \"certified-operators-m8kv6\" (UID: \"a098079d-37fa-4f22-a9aa-ee57589c52ab\") " pod="openshift-marketplace/certified-operators-m8kv6" Mar 21 08:53:50 crc kubenswrapper[4696]: I0321 08:53:50.993678 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a098079d-37fa-4f22-a9aa-ee57589c52ab-catalog-content\") pod \"certified-operators-m8kv6\" (UID: \"a098079d-37fa-4f22-a9aa-ee57589c52ab\") " pod="openshift-marketplace/certified-operators-m8kv6" Mar 21 08:53:51 crc kubenswrapper[4696]: I0321 08:53:51.015414 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pz8v\" (UniqueName: \"kubernetes.io/projected/a098079d-37fa-4f22-a9aa-ee57589c52ab-kube-api-access-8pz8v\") pod \"certified-operators-m8kv6\" (UID: \"a098079d-37fa-4f22-a9aa-ee57589c52ab\") " pod="openshift-marketplace/certified-operators-m8kv6" Mar 21 08:53:51 crc kubenswrapper[4696]: I0321 08:53:51.215364 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m8kv6" Mar 21 08:53:51 crc kubenswrapper[4696]: I0321 08:53:51.744910 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m8kv6"] Mar 21 08:53:52 crc kubenswrapper[4696]: I0321 08:53:52.195792 4696 generic.go:334] "Generic (PLEG): container finished" podID="a098079d-37fa-4f22-a9aa-ee57589c52ab" containerID="5f871eadabd03ff40252d1a381a70c94676a54cebb05c338725da55a34962a7d" exitCode=0 Mar 21 08:53:52 crc kubenswrapper[4696]: I0321 08:53:52.195864 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m8kv6" event={"ID":"a098079d-37fa-4f22-a9aa-ee57589c52ab","Type":"ContainerDied","Data":"5f871eadabd03ff40252d1a381a70c94676a54cebb05c338725da55a34962a7d"} Mar 21 08:53:52 crc kubenswrapper[4696]: I0321 08:53:52.196092 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m8kv6" event={"ID":"a098079d-37fa-4f22-a9aa-ee57589c52ab","Type":"ContainerStarted","Data":"0a4ba02e3fef53b0c2d05fadc74bfdb43e3a63542ca4ddbb9962da5a6fde7953"} Mar 21 08:53:53 crc kubenswrapper[4696]: I0321 08:53:53.217859 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m8kv6" event={"ID":"a098079d-37fa-4f22-a9aa-ee57589c52ab","Type":"ContainerStarted","Data":"b48b2b672714aea4b586025209bf0cf86c6a70aea4e368d862d4427d04c0b9af"} Mar 21 08:53:55 crc kubenswrapper[4696]: I0321 08:53:55.239389 4696 generic.go:334] "Generic (PLEG): container finished" podID="a098079d-37fa-4f22-a9aa-ee57589c52ab" containerID="b48b2b672714aea4b586025209bf0cf86c6a70aea4e368d862d4427d04c0b9af" exitCode=0 Mar 21 08:53:55 crc kubenswrapper[4696]: I0321 08:53:55.239445 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m8kv6" event={"ID":"a098079d-37fa-4f22-a9aa-ee57589c52ab","Type":"ContainerDied","Data":"b48b2b672714aea4b586025209bf0cf86c6a70aea4e368d862d4427d04c0b9af"} Mar 21 08:53:56 crc kubenswrapper[4696]: I0321 08:53:56.254480 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m8kv6" event={"ID":"a098079d-37fa-4f22-a9aa-ee57589c52ab","Type":"ContainerStarted","Data":"9ee6e3ef84a46e65e2c4b7c8ebae6954a056fb1f3b16486f30ab2c37c664073e"} Mar 21 08:53:56 crc kubenswrapper[4696]: I0321 08:53:56.281456 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m8kv6" podStartSLOduration=2.786299255 podStartE2EDuration="6.281431672s" podCreationTimestamp="2026-03-21 08:53:50 +0000 UTC" firstStartedPulling="2026-03-21 08:53:52.197880562 +0000 UTC m=+1566.318761275" lastFinishedPulling="2026-03-21 08:53:55.693012979 +0000 UTC m=+1569.813893692" observedRunningTime="2026-03-21 08:53:56.271417161 +0000 UTC m=+1570.392297864" watchObservedRunningTime="2026-03-21 08:53:56.281431672 +0000 UTC m=+1570.402312385" Mar 21 08:53:58 crc kubenswrapper[4696]: I0321 08:53:58.257315 4696 scope.go:117] "RemoveContainer" containerID="02f71b56389f4f65c6fb6ec54074e2adfea253d692cbac199e64991a0696ee68" Mar 21 08:53:58 crc kubenswrapper[4696]: I0321 08:53:58.299903 4696 scope.go:117] "RemoveContainer" containerID="1c5e7e384295a1621158846edea9c769dbc91bd01ee2e2325ad8211be97cf9e6" Mar 21 08:53:58 crc kubenswrapper[4696]: I0321 08:53:58.327015 4696 scope.go:117] "RemoveContainer" containerID="a95ef0ed856984fc1e9f6ad813bed8db93b34bf18988bcd1f080a5078c1422c2" Mar 21 08:53:58 crc kubenswrapper[4696]: I0321 08:53:58.353586 4696 scope.go:117] "RemoveContainer" containerID="d6a2c9083ae133bb5944b16dc225cccbf9e3a5f9ea56291a7f5dd48fc657a5ad" Mar 21 08:53:58 crc kubenswrapper[4696]: I0321 08:53:58.397133 4696 scope.go:117] "RemoveContainer" containerID="027b0a845d9b25edaa4aec6a90d1a489fea68982c21b4e5196545686f124576a" Mar 21 08:53:58 crc kubenswrapper[4696]: I0321 08:53:58.496297 4696 scope.go:117] "RemoveContainer" containerID="17745124f29da5ac840d31851871596d27883037e3f61a184b8f34cb87fd4a9b" Mar 21 08:53:58 crc kubenswrapper[4696]: I0321 08:53:58.525568 4696 scope.go:117] "RemoveContainer" containerID="83e39ef69714631ade43ac7b3e551d20a5d282e9d803ae8961e6ae213c1e452e" Mar 21 08:53:58 crc kubenswrapper[4696]: I0321 08:53:58.559738 4696 scope.go:117] "RemoveContainer" containerID="34604a8d20dd45769c938fc4e612566bd5eb98466fd47a1560d6276fec96a001" Mar 21 08:53:58 crc kubenswrapper[4696]: I0321 08:53:58.597419 4696 scope.go:117] "RemoveContainer" containerID="5f83ec04b39227800d19bc750797e0e8b8ffaaee0a89e3f3298d96b0a49de970" Mar 21 08:53:58 crc kubenswrapper[4696]: I0321 08:53:58.618612 4696 scope.go:117] "RemoveContainer" containerID="f8977d91c22cb7cd0bc89772b38b637d0532cf700af1c078f99e5082fae85989" Mar 21 08:53:58 crc kubenswrapper[4696]: I0321 08:53:58.638467 4696 scope.go:117] "RemoveContainer" containerID="3ea87006f13706ba3d8ac0954309bbbc6e34e6fc1350bd3952ec2ae56578a677" Mar 21 08:54:00 crc kubenswrapper[4696]: I0321 08:54:00.151153 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568054-l85jh"] Mar 21 08:54:00 crc kubenswrapper[4696]: I0321 08:54:00.164461 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568054-l85jh"] Mar 21 08:54:00 crc kubenswrapper[4696]: I0321 08:54:00.164574 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568054-l85jh" Mar 21 08:54:00 crc kubenswrapper[4696]: I0321 08:54:00.169811 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 08:54:00 crc kubenswrapper[4696]: I0321 08:54:00.170120 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 08:54:00 crc kubenswrapper[4696]: I0321 08:54:00.170231 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 08:54:00 crc kubenswrapper[4696]: I0321 08:54:00.293527 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chljj\" (UniqueName: \"kubernetes.io/projected/03261136-a65c-455b-9e7d-f24d95e68322-kube-api-access-chljj\") pod \"auto-csr-approver-29568054-l85jh\" (UID: \"03261136-a65c-455b-9e7d-f24d95e68322\") " pod="openshift-infra/auto-csr-approver-29568054-l85jh" Mar 21 08:54:00 crc kubenswrapper[4696]: I0321 08:54:00.341243 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:54:00 crc kubenswrapper[4696]: I0321 08:54:00.341298 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:54:00 crc kubenswrapper[4696]: I0321 08:54:00.395144 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chljj\" (UniqueName: \"kubernetes.io/projected/03261136-a65c-455b-9e7d-f24d95e68322-kube-api-access-chljj\") pod \"auto-csr-approver-29568054-l85jh\" (UID: \"03261136-a65c-455b-9e7d-f24d95e68322\") " pod="openshift-infra/auto-csr-approver-29568054-l85jh" Mar 21 08:54:00 crc kubenswrapper[4696]: I0321 08:54:00.423450 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chljj\" (UniqueName: \"kubernetes.io/projected/03261136-a65c-455b-9e7d-f24d95e68322-kube-api-access-chljj\") pod \"auto-csr-approver-29568054-l85jh\" (UID: \"03261136-a65c-455b-9e7d-f24d95e68322\") " pod="openshift-infra/auto-csr-approver-29568054-l85jh" Mar 21 08:54:00 crc kubenswrapper[4696]: I0321 08:54:00.529259 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568054-l85jh" Mar 21 08:54:01 crc kubenswrapper[4696]: I0321 08:54:01.024699 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568054-l85jh"] Mar 21 08:54:01 crc kubenswrapper[4696]: I0321 08:54:01.215809 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m8kv6" Mar 21 08:54:01 crc kubenswrapper[4696]: I0321 08:54:01.215890 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m8kv6" Mar 21 08:54:01 crc kubenswrapper[4696]: I0321 08:54:01.306269 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568054-l85jh" event={"ID":"03261136-a65c-455b-9e7d-f24d95e68322","Type":"ContainerStarted","Data":"de00de8f1a27298ffd2f16e0c94090f7d215a793d1bbc22e028d2c28356c5dee"} Mar 21 08:54:02 crc kubenswrapper[4696]: I0321 08:54:02.264270 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-m8kv6" podUID="a098079d-37fa-4f22-a9aa-ee57589c52ab" containerName="registry-server" probeResult="failure" output=< Mar 21 08:54:02 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Mar 21 08:54:02 crc kubenswrapper[4696]: > Mar 21 08:54:02 crc kubenswrapper[4696]: I0321 08:54:02.320444 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568054-l85jh" event={"ID":"03261136-a65c-455b-9e7d-f24d95e68322","Type":"ContainerStarted","Data":"5c2f857bc27ae26eded78f4c0167984e206a5936f4a62a234f63d4edb7f3b89f"} Mar 21 08:54:02 crc kubenswrapper[4696]: I0321 08:54:02.338575 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29568054-l85jh" podStartSLOduration=1.528042234 podStartE2EDuration="2.338560164s" podCreationTimestamp="2026-03-21 08:54:00 +0000 UTC" firstStartedPulling="2026-03-21 08:54:01.030985246 +0000 UTC m=+1575.151865959" lastFinishedPulling="2026-03-21 08:54:01.841503176 +0000 UTC m=+1575.962383889" observedRunningTime="2026-03-21 08:54:02.335999423 +0000 UTC m=+1576.456880156" watchObservedRunningTime="2026-03-21 08:54:02.338560164 +0000 UTC m=+1576.459440877" Mar 21 08:54:03 crc kubenswrapper[4696]: I0321 08:54:03.332525 4696 generic.go:334] "Generic (PLEG): container finished" podID="03261136-a65c-455b-9e7d-f24d95e68322" containerID="5c2f857bc27ae26eded78f4c0167984e206a5936f4a62a234f63d4edb7f3b89f" exitCode=0 Mar 21 08:54:03 crc kubenswrapper[4696]: I0321 08:54:03.332729 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568054-l85jh" event={"ID":"03261136-a65c-455b-9e7d-f24d95e68322","Type":"ContainerDied","Data":"5c2f857bc27ae26eded78f4c0167984e206a5936f4a62a234f63d4edb7f3b89f"} Mar 21 08:54:05 crc kubenswrapper[4696]: I0321 08:54:05.085623 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568054-l85jh" Mar 21 08:54:05 crc kubenswrapper[4696]: I0321 08:54:05.133548 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chljj\" (UniqueName: \"kubernetes.io/projected/03261136-a65c-455b-9e7d-f24d95e68322-kube-api-access-chljj\") pod \"03261136-a65c-455b-9e7d-f24d95e68322\" (UID: \"03261136-a65c-455b-9e7d-f24d95e68322\") " Mar 21 08:54:05 crc kubenswrapper[4696]: I0321 08:54:05.142118 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03261136-a65c-455b-9e7d-f24d95e68322-kube-api-access-chljj" (OuterVolumeSpecName: "kube-api-access-chljj") pod "03261136-a65c-455b-9e7d-f24d95e68322" (UID: "03261136-a65c-455b-9e7d-f24d95e68322"). InnerVolumeSpecName "kube-api-access-chljj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:54:05 crc kubenswrapper[4696]: I0321 08:54:05.236694 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chljj\" (UniqueName: \"kubernetes.io/projected/03261136-a65c-455b-9e7d-f24d95e68322-kube-api-access-chljj\") on node \"crc\" DevicePath \"\"" Mar 21 08:54:05 crc kubenswrapper[4696]: I0321 08:54:05.351549 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568054-l85jh" event={"ID":"03261136-a65c-455b-9e7d-f24d95e68322","Type":"ContainerDied","Data":"de00de8f1a27298ffd2f16e0c94090f7d215a793d1bbc22e028d2c28356c5dee"} Mar 21 08:54:05 crc kubenswrapper[4696]: I0321 08:54:05.351596 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de00de8f1a27298ffd2f16e0c94090f7d215a793d1bbc22e028d2c28356c5dee" Mar 21 08:54:05 crc kubenswrapper[4696]: I0321 08:54:05.351650 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568054-l85jh" Mar 21 08:54:05 crc kubenswrapper[4696]: I0321 08:54:05.412438 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568048-j276n"] Mar 21 08:54:05 crc kubenswrapper[4696]: I0321 08:54:05.421719 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568048-j276n"] Mar 21 08:54:06 crc kubenswrapper[4696]: I0321 08:54:06.549540 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41361f63-e76c-42d2-ae18-32c5e772773a" path="/var/lib/kubelet/pods/41361f63-e76c-42d2-ae18-32c5e772773a/volumes" Mar 21 08:54:11 crc kubenswrapper[4696]: I0321 08:54:11.279136 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m8kv6" Mar 21 08:54:11 crc kubenswrapper[4696]: I0321 08:54:11.335908 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m8kv6" Mar 21 08:54:11 crc kubenswrapper[4696]: I0321 08:54:11.515342 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m8kv6"] Mar 21 08:54:12 crc kubenswrapper[4696]: I0321 08:54:12.450299 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-m8kv6" podUID="a098079d-37fa-4f22-a9aa-ee57589c52ab" containerName="registry-server" containerID="cri-o://9ee6e3ef84a46e65e2c4b7c8ebae6954a056fb1f3b16486f30ab2c37c664073e" gracePeriod=2 Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.410891 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m8kv6" Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.464736 4696 generic.go:334] "Generic (PLEG): container finished" podID="a098079d-37fa-4f22-a9aa-ee57589c52ab" containerID="9ee6e3ef84a46e65e2c4b7c8ebae6954a056fb1f3b16486f30ab2c37c664073e" exitCode=0 Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.464784 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m8kv6" event={"ID":"a098079d-37fa-4f22-a9aa-ee57589c52ab","Type":"ContainerDied","Data":"9ee6e3ef84a46e65e2c4b7c8ebae6954a056fb1f3b16486f30ab2c37c664073e"} Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.464819 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m8kv6" event={"ID":"a098079d-37fa-4f22-a9aa-ee57589c52ab","Type":"ContainerDied","Data":"0a4ba02e3fef53b0c2d05fadc74bfdb43e3a63542ca4ddbb9962da5a6fde7953"} Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.464859 4696 scope.go:117] "RemoveContainer" containerID="9ee6e3ef84a46e65e2c4b7c8ebae6954a056fb1f3b16486f30ab2c37c664073e" Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.465145 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m8kv6" Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.504016 4696 scope.go:117] "RemoveContainer" containerID="b48b2b672714aea4b586025209bf0cf86c6a70aea4e368d862d4427d04c0b9af" Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.513011 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a098079d-37fa-4f22-a9aa-ee57589c52ab-catalog-content\") pod \"a098079d-37fa-4f22-a9aa-ee57589c52ab\" (UID: \"a098079d-37fa-4f22-a9aa-ee57589c52ab\") " Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.513112 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a098079d-37fa-4f22-a9aa-ee57589c52ab-utilities\") pod \"a098079d-37fa-4f22-a9aa-ee57589c52ab\" (UID: \"a098079d-37fa-4f22-a9aa-ee57589c52ab\") " Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.513195 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pz8v\" (UniqueName: \"kubernetes.io/projected/a098079d-37fa-4f22-a9aa-ee57589c52ab-kube-api-access-8pz8v\") pod \"a098079d-37fa-4f22-a9aa-ee57589c52ab\" (UID: \"a098079d-37fa-4f22-a9aa-ee57589c52ab\") " Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.514512 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a098079d-37fa-4f22-a9aa-ee57589c52ab-utilities" (OuterVolumeSpecName: "utilities") pod "a098079d-37fa-4f22-a9aa-ee57589c52ab" (UID: "a098079d-37fa-4f22-a9aa-ee57589c52ab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.519310 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a098079d-37fa-4f22-a9aa-ee57589c52ab-kube-api-access-8pz8v" (OuterVolumeSpecName: "kube-api-access-8pz8v") pod "a098079d-37fa-4f22-a9aa-ee57589c52ab" (UID: "a098079d-37fa-4f22-a9aa-ee57589c52ab"). InnerVolumeSpecName "kube-api-access-8pz8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.526017 4696 scope.go:117] "RemoveContainer" containerID="5f871eadabd03ff40252d1a381a70c94676a54cebb05c338725da55a34962a7d" Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.575037 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a098079d-37fa-4f22-a9aa-ee57589c52ab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a098079d-37fa-4f22-a9aa-ee57589c52ab" (UID: "a098079d-37fa-4f22-a9aa-ee57589c52ab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.615520 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a098079d-37fa-4f22-a9aa-ee57589c52ab-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.615674 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a098079d-37fa-4f22-a9aa-ee57589c52ab-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.615897 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pz8v\" (UniqueName: \"kubernetes.io/projected/a098079d-37fa-4f22-a9aa-ee57589c52ab-kube-api-access-8pz8v\") on node \"crc\" DevicePath \"\"" Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.636104 4696 scope.go:117] "RemoveContainer" containerID="9ee6e3ef84a46e65e2c4b7c8ebae6954a056fb1f3b16486f30ab2c37c664073e" Mar 21 08:54:13 crc kubenswrapper[4696]: E0321 08:54:13.636561 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ee6e3ef84a46e65e2c4b7c8ebae6954a056fb1f3b16486f30ab2c37c664073e\": container with ID starting with 9ee6e3ef84a46e65e2c4b7c8ebae6954a056fb1f3b16486f30ab2c37c664073e not found: ID does not exist" containerID="9ee6e3ef84a46e65e2c4b7c8ebae6954a056fb1f3b16486f30ab2c37c664073e" Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.636595 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ee6e3ef84a46e65e2c4b7c8ebae6954a056fb1f3b16486f30ab2c37c664073e"} err="failed to get container status \"9ee6e3ef84a46e65e2c4b7c8ebae6954a056fb1f3b16486f30ab2c37c664073e\": rpc error: code = NotFound desc = could not find container \"9ee6e3ef84a46e65e2c4b7c8ebae6954a056fb1f3b16486f30ab2c37c664073e\": container with ID starting with 9ee6e3ef84a46e65e2c4b7c8ebae6954a056fb1f3b16486f30ab2c37c664073e not found: ID does not exist" Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.636622 4696 scope.go:117] "RemoveContainer" containerID="b48b2b672714aea4b586025209bf0cf86c6a70aea4e368d862d4427d04c0b9af" Mar 21 08:54:13 crc kubenswrapper[4696]: E0321 08:54:13.637903 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b48b2b672714aea4b586025209bf0cf86c6a70aea4e368d862d4427d04c0b9af\": container with ID starting with b48b2b672714aea4b586025209bf0cf86c6a70aea4e368d862d4427d04c0b9af not found: ID does not exist" containerID="b48b2b672714aea4b586025209bf0cf86c6a70aea4e368d862d4427d04c0b9af" Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.638024 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b48b2b672714aea4b586025209bf0cf86c6a70aea4e368d862d4427d04c0b9af"} err="failed to get container status \"b48b2b672714aea4b586025209bf0cf86c6a70aea4e368d862d4427d04c0b9af\": rpc error: code = NotFound desc = could not find container \"b48b2b672714aea4b586025209bf0cf86c6a70aea4e368d862d4427d04c0b9af\": container with ID starting with b48b2b672714aea4b586025209bf0cf86c6a70aea4e368d862d4427d04c0b9af not found: ID does not exist" Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.638147 4696 scope.go:117] "RemoveContainer" containerID="5f871eadabd03ff40252d1a381a70c94676a54cebb05c338725da55a34962a7d" Mar 21 08:54:13 crc kubenswrapper[4696]: E0321 08:54:13.640430 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f871eadabd03ff40252d1a381a70c94676a54cebb05c338725da55a34962a7d\": container with ID starting with 5f871eadabd03ff40252d1a381a70c94676a54cebb05c338725da55a34962a7d not found: ID does not exist" containerID="5f871eadabd03ff40252d1a381a70c94676a54cebb05c338725da55a34962a7d" Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.640540 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f871eadabd03ff40252d1a381a70c94676a54cebb05c338725da55a34962a7d"} err="failed to get container status \"5f871eadabd03ff40252d1a381a70c94676a54cebb05c338725da55a34962a7d\": rpc error: code = NotFound desc = could not find container \"5f871eadabd03ff40252d1a381a70c94676a54cebb05c338725da55a34962a7d\": container with ID starting with 5f871eadabd03ff40252d1a381a70c94676a54cebb05c338725da55a34962a7d not found: ID does not exist" Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.831061 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m8kv6"] Mar 21 08:54:13 crc kubenswrapper[4696]: I0321 08:54:13.843579 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-m8kv6"] Mar 21 08:54:14 crc kubenswrapper[4696]: I0321 08:54:14.546041 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a098079d-37fa-4f22-a9aa-ee57589c52ab" path="/var/lib/kubelet/pods/a098079d-37fa-4f22-a9aa-ee57589c52ab/volumes" Mar 21 08:54:30 crc kubenswrapper[4696]: I0321 08:54:30.341548 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:54:30 crc kubenswrapper[4696]: I0321 08:54:30.342073 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:54:58 crc kubenswrapper[4696]: I0321 08:54:58.921121 4696 scope.go:117] "RemoveContainer" containerID="b296773d278d198bdcff27110fa8b8aee8a0160dbe3d66d891888e4b8c7a2532" Mar 21 08:55:00 crc kubenswrapper[4696]: I0321 08:55:00.341117 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 08:55:00 crc kubenswrapper[4696]: I0321 08:55:00.342203 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 08:55:00 crc kubenswrapper[4696]: I0321 08:55:00.342340 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 08:55:00 crc kubenswrapper[4696]: I0321 08:55:00.343227 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959"} pod="openshift-machine-config-operator/machine-config-daemon-z7srw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 08:55:00 crc kubenswrapper[4696]: I0321 08:55:00.343403 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" containerID="cri-o://a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" gracePeriod=600 Mar 21 08:55:00 crc kubenswrapper[4696]: E0321 08:55:00.471880 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:55:00 crc kubenswrapper[4696]: I0321 08:55:00.938446 4696 generic.go:334] "Generic (PLEG): container finished" podID="daaf227a-2305-495e-8495-a280abcd8e10" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" exitCode=0 Mar 21 08:55:00 crc kubenswrapper[4696]: I0321 08:55:00.938488 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerDied","Data":"a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959"} Mar 21 08:55:00 crc kubenswrapper[4696]: I0321 08:55:00.938552 4696 scope.go:117] "RemoveContainer" containerID="25b5c4ade4d0fb1258c68b7e200638ed962f1ba1ce964274a9e7589cd0a163c2" Mar 21 08:55:00 crc kubenswrapper[4696]: I0321 08:55:00.939390 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:55:00 crc kubenswrapper[4696]: E0321 08:55:00.939651 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:55:16 crc kubenswrapper[4696]: I0321 08:55:16.542581 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:55:16 crc kubenswrapper[4696]: E0321 08:55:16.543390 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:55:27 crc kubenswrapper[4696]: I0321 08:55:27.535137 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:55:27 crc kubenswrapper[4696]: E0321 08:55:27.535900 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:55:33 crc kubenswrapper[4696]: I0321 08:55:33.279962 4696 generic.go:334] "Generic (PLEG): container finished" podID="875486e6-3369-4601-bc53-00c2aaf85707" containerID="1d2e2ff8b196017f4c18a9283358e136c6ea0eab6e1251edb32e8f96e2e90cf3" exitCode=0 Mar 21 08:55:33 crc kubenswrapper[4696]: I0321 08:55:33.280552 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq" event={"ID":"875486e6-3369-4601-bc53-00c2aaf85707","Type":"ContainerDied","Data":"1d2e2ff8b196017f4c18a9283358e136c6ea0eab6e1251edb32e8f96e2e90cf3"} Mar 21 08:55:35 crc kubenswrapper[4696]: I0321 08:55:35.300783 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq" event={"ID":"875486e6-3369-4601-bc53-00c2aaf85707","Type":"ContainerDied","Data":"811538461d3743dc98dea22cdf84f6a5ea72e332c1f86863bf87582bd85c4617"} Mar 21 08:55:35 crc kubenswrapper[4696]: I0321 08:55:35.301103 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="811538461d3743dc98dea22cdf84f6a5ea72e332c1f86863bf87582bd85c4617" Mar 21 08:55:35 crc kubenswrapper[4696]: I0321 08:55:35.370654 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq" Mar 21 08:55:35 crc kubenswrapper[4696]: I0321 08:55:35.566500 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/875486e6-3369-4601-bc53-00c2aaf85707-inventory\") pod \"875486e6-3369-4601-bc53-00c2aaf85707\" (UID: \"875486e6-3369-4601-bc53-00c2aaf85707\") " Mar 21 08:55:35 crc kubenswrapper[4696]: I0321 08:55:35.566925 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6pl2\" (UniqueName: \"kubernetes.io/projected/875486e6-3369-4601-bc53-00c2aaf85707-kube-api-access-s6pl2\") pod \"875486e6-3369-4601-bc53-00c2aaf85707\" (UID: \"875486e6-3369-4601-bc53-00c2aaf85707\") " Mar 21 08:55:35 crc kubenswrapper[4696]: I0321 08:55:35.567066 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/875486e6-3369-4601-bc53-00c2aaf85707-bootstrap-combined-ca-bundle\") pod \"875486e6-3369-4601-bc53-00c2aaf85707\" (UID: \"875486e6-3369-4601-bc53-00c2aaf85707\") " Mar 21 08:55:35 crc kubenswrapper[4696]: I0321 08:55:35.567258 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/875486e6-3369-4601-bc53-00c2aaf85707-ssh-key-openstack-edpm-ipam\") pod \"875486e6-3369-4601-bc53-00c2aaf85707\" (UID: \"875486e6-3369-4601-bc53-00c2aaf85707\") " Mar 21 08:55:35 crc kubenswrapper[4696]: I0321 08:55:35.572024 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/875486e6-3369-4601-bc53-00c2aaf85707-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "875486e6-3369-4601-bc53-00c2aaf85707" (UID: "875486e6-3369-4601-bc53-00c2aaf85707"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:55:35 crc kubenswrapper[4696]: I0321 08:55:35.583300 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/875486e6-3369-4601-bc53-00c2aaf85707-kube-api-access-s6pl2" (OuterVolumeSpecName: "kube-api-access-s6pl2") pod "875486e6-3369-4601-bc53-00c2aaf85707" (UID: "875486e6-3369-4601-bc53-00c2aaf85707"). InnerVolumeSpecName "kube-api-access-s6pl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:55:35 crc kubenswrapper[4696]: I0321 08:55:35.597179 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/875486e6-3369-4601-bc53-00c2aaf85707-inventory" (OuterVolumeSpecName: "inventory") pod "875486e6-3369-4601-bc53-00c2aaf85707" (UID: "875486e6-3369-4601-bc53-00c2aaf85707"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:55:35 crc kubenswrapper[4696]: I0321 08:55:35.609058 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/875486e6-3369-4601-bc53-00c2aaf85707-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "875486e6-3369-4601-bc53-00c2aaf85707" (UID: "875486e6-3369-4601-bc53-00c2aaf85707"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:55:35 crc kubenswrapper[4696]: I0321 08:55:35.670546 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6pl2\" (UniqueName: \"kubernetes.io/projected/875486e6-3369-4601-bc53-00c2aaf85707-kube-api-access-s6pl2\") on node \"crc\" DevicePath \"\"" Mar 21 08:55:35 crc kubenswrapper[4696]: I0321 08:55:35.670592 4696 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/875486e6-3369-4601-bc53-00c2aaf85707-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 08:55:35 crc kubenswrapper[4696]: I0321 08:55:35.670606 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/875486e6-3369-4601-bc53-00c2aaf85707-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 08:55:35 crc kubenswrapper[4696]: I0321 08:55:35.670619 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/875486e6-3369-4601-bc53-00c2aaf85707-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.309874 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.469068 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zncdp"] Mar 21 08:55:36 crc kubenswrapper[4696]: E0321 08:55:36.469519 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a098079d-37fa-4f22-a9aa-ee57589c52ab" containerName="registry-server" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.469535 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="a098079d-37fa-4f22-a9aa-ee57589c52ab" containerName="registry-server" Mar 21 08:55:36 crc kubenswrapper[4696]: E0321 08:55:36.469564 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="875486e6-3369-4601-bc53-00c2aaf85707" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.469572 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="875486e6-3369-4601-bc53-00c2aaf85707" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 21 08:55:36 crc kubenswrapper[4696]: E0321 08:55:36.469590 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03261136-a65c-455b-9e7d-f24d95e68322" containerName="oc" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.469596 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="03261136-a65c-455b-9e7d-f24d95e68322" containerName="oc" Mar 21 08:55:36 crc kubenswrapper[4696]: E0321 08:55:36.469625 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a098079d-37fa-4f22-a9aa-ee57589c52ab" containerName="extract-utilities" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.469632 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="a098079d-37fa-4f22-a9aa-ee57589c52ab" containerName="extract-utilities" Mar 21 08:55:36 crc kubenswrapper[4696]: E0321 08:55:36.469648 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a098079d-37fa-4f22-a9aa-ee57589c52ab" containerName="extract-content" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.469653 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="a098079d-37fa-4f22-a9aa-ee57589c52ab" containerName="extract-content" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.469870 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="03261136-a65c-455b-9e7d-f24d95e68322" containerName="oc" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.469886 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="a098079d-37fa-4f22-a9aa-ee57589c52ab" containerName="registry-server" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.469913 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="875486e6-3369-4601-bc53-00c2aaf85707" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.470676 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zncdp" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.472606 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.472779 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.473130 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bqp9r" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.473377 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.510489 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zncdp"] Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.587001 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c756d18c-c925-44bd-911f-0b82d961c66e-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zncdp\" (UID: \"c756d18c-c925-44bd-911f-0b82d961c66e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zncdp" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.587344 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c756d18c-c925-44bd-911f-0b82d961c66e-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zncdp\" (UID: \"c756d18c-c925-44bd-911f-0b82d961c66e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zncdp" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.587421 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8hxt\" (UniqueName: \"kubernetes.io/projected/c756d18c-c925-44bd-911f-0b82d961c66e-kube-api-access-n8hxt\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zncdp\" (UID: \"c756d18c-c925-44bd-911f-0b82d961c66e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zncdp" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.689671 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c756d18c-c925-44bd-911f-0b82d961c66e-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zncdp\" (UID: \"c756d18c-c925-44bd-911f-0b82d961c66e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zncdp" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.689724 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8hxt\" (UniqueName: \"kubernetes.io/projected/c756d18c-c925-44bd-911f-0b82d961c66e-kube-api-access-n8hxt\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zncdp\" (UID: \"c756d18c-c925-44bd-911f-0b82d961c66e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zncdp" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.689800 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c756d18c-c925-44bd-911f-0b82d961c66e-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zncdp\" (UID: \"c756d18c-c925-44bd-911f-0b82d961c66e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zncdp" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.694074 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c756d18c-c925-44bd-911f-0b82d961c66e-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zncdp\" (UID: \"c756d18c-c925-44bd-911f-0b82d961c66e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zncdp" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.694625 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c756d18c-c925-44bd-911f-0b82d961c66e-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zncdp\" (UID: \"c756d18c-c925-44bd-911f-0b82d961c66e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zncdp" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.707686 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8hxt\" (UniqueName: \"kubernetes.io/projected/c756d18c-c925-44bd-911f-0b82d961c66e-kube-api-access-n8hxt\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zncdp\" (UID: \"c756d18c-c925-44bd-911f-0b82d961c66e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zncdp" Mar 21 08:55:36 crc kubenswrapper[4696]: I0321 08:55:36.810527 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zncdp" Mar 21 08:55:37 crc kubenswrapper[4696]: I0321 08:55:37.369210 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zncdp"] Mar 21 08:55:38 crc kubenswrapper[4696]: I0321 08:55:38.330738 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zncdp" event={"ID":"c756d18c-c925-44bd-911f-0b82d961c66e","Type":"ContainerStarted","Data":"37990705b78f5cf85a469b6c4ccc626bee25e36ae7fda2c36575e1283f2f1bf7"} Mar 21 08:55:38 crc kubenswrapper[4696]: I0321 08:55:38.331248 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zncdp" event={"ID":"c756d18c-c925-44bd-911f-0b82d961c66e","Type":"ContainerStarted","Data":"7de3e383f017f3451359a7ea99d48ef414403843f1833f29bfc0f7dac18856c1"} Mar 21 08:55:38 crc kubenswrapper[4696]: I0321 08:55:38.352723 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zncdp" podStartSLOduration=1.921472775 podStartE2EDuration="2.352703113s" podCreationTimestamp="2026-03-21 08:55:36 +0000 UTC" firstStartedPulling="2026-03-21 08:55:37.377730922 +0000 UTC m=+1671.498611635" lastFinishedPulling="2026-03-21 08:55:37.80896125 +0000 UTC m=+1671.929841973" observedRunningTime="2026-03-21 08:55:38.346351719 +0000 UTC m=+1672.467232442" watchObservedRunningTime="2026-03-21 08:55:38.352703113 +0000 UTC m=+1672.473583826" Mar 21 08:55:41 crc kubenswrapper[4696]: I0321 08:55:41.535075 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:55:41 crc kubenswrapper[4696]: E0321 08:55:41.535739 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:55:56 crc kubenswrapper[4696]: I0321 08:55:56.542729 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:55:56 crc kubenswrapper[4696]: E0321 08:55:56.543685 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:56:00 crc kubenswrapper[4696]: I0321 08:56:00.148966 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568056-sgpwk"] Mar 21 08:56:00 crc kubenswrapper[4696]: I0321 08:56:00.150832 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568056-sgpwk" Mar 21 08:56:00 crc kubenswrapper[4696]: I0321 08:56:00.154062 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 08:56:00 crc kubenswrapper[4696]: I0321 08:56:00.157378 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 08:56:00 crc kubenswrapper[4696]: I0321 08:56:00.158651 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 08:56:00 crc kubenswrapper[4696]: I0321 08:56:00.177616 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568056-sgpwk"] Mar 21 08:56:00 crc kubenswrapper[4696]: I0321 08:56:00.335994 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xztc8\" (UniqueName: \"kubernetes.io/projected/f6936cdd-5be6-4162-a976-47d5df686290-kube-api-access-xztc8\") pod \"auto-csr-approver-29568056-sgpwk\" (UID: \"f6936cdd-5be6-4162-a976-47d5df686290\") " pod="openshift-infra/auto-csr-approver-29568056-sgpwk" Mar 21 08:56:00 crc kubenswrapper[4696]: I0321 08:56:00.438628 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xztc8\" (UniqueName: \"kubernetes.io/projected/f6936cdd-5be6-4162-a976-47d5df686290-kube-api-access-xztc8\") pod \"auto-csr-approver-29568056-sgpwk\" (UID: \"f6936cdd-5be6-4162-a976-47d5df686290\") " pod="openshift-infra/auto-csr-approver-29568056-sgpwk" Mar 21 08:56:00 crc kubenswrapper[4696]: I0321 08:56:00.460525 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xztc8\" (UniqueName: \"kubernetes.io/projected/f6936cdd-5be6-4162-a976-47d5df686290-kube-api-access-xztc8\") pod \"auto-csr-approver-29568056-sgpwk\" (UID: \"f6936cdd-5be6-4162-a976-47d5df686290\") " pod="openshift-infra/auto-csr-approver-29568056-sgpwk" Mar 21 08:56:00 crc kubenswrapper[4696]: I0321 08:56:00.489710 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568056-sgpwk" Mar 21 08:56:01 crc kubenswrapper[4696]: I0321 08:56:01.214272 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568056-sgpwk"] Mar 21 08:56:01 crc kubenswrapper[4696]: W0321 08:56:01.218256 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6936cdd_5be6_4162_a976_47d5df686290.slice/crio-8428f189f45f3e6479de883ed0e5eb1cc3bd18912e424b389d9a6892cfdf08ca WatchSource:0}: Error finding container 8428f189f45f3e6479de883ed0e5eb1cc3bd18912e424b389d9a6892cfdf08ca: Status 404 returned error can't find the container with id 8428f189f45f3e6479de883ed0e5eb1cc3bd18912e424b389d9a6892cfdf08ca Mar 21 08:56:01 crc kubenswrapper[4696]: I0321 08:56:01.591389 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568056-sgpwk" event={"ID":"f6936cdd-5be6-4162-a976-47d5df686290","Type":"ContainerStarted","Data":"8428f189f45f3e6479de883ed0e5eb1cc3bd18912e424b389d9a6892cfdf08ca"} Mar 21 08:56:02 crc kubenswrapper[4696]: I0321 08:56:02.603125 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568056-sgpwk" event={"ID":"f6936cdd-5be6-4162-a976-47d5df686290","Type":"ContainerStarted","Data":"ad43bf346eb6430191fea0d17617da01fb6e57eca827d0006fe34f07eaa54a53"} Mar 21 08:56:02 crc kubenswrapper[4696]: I0321 08:56:02.627899 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29568056-sgpwk" podStartSLOduration=1.750136458 podStartE2EDuration="2.627879365s" podCreationTimestamp="2026-03-21 08:56:00 +0000 UTC" firstStartedPulling="2026-03-21 08:56:01.221394918 +0000 UTC m=+1695.342275631" lastFinishedPulling="2026-03-21 08:56:02.099137825 +0000 UTC m=+1696.220018538" observedRunningTime="2026-03-21 08:56:02.616771801 +0000 UTC m=+1696.737652504" watchObservedRunningTime="2026-03-21 08:56:02.627879365 +0000 UTC m=+1696.748760078" Mar 21 08:56:03 crc kubenswrapper[4696]: I0321 08:56:03.614732 4696 generic.go:334] "Generic (PLEG): container finished" podID="f6936cdd-5be6-4162-a976-47d5df686290" containerID="ad43bf346eb6430191fea0d17617da01fb6e57eca827d0006fe34f07eaa54a53" exitCode=0 Mar 21 08:56:03 crc kubenswrapper[4696]: I0321 08:56:03.614796 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568056-sgpwk" event={"ID":"f6936cdd-5be6-4162-a976-47d5df686290","Type":"ContainerDied","Data":"ad43bf346eb6430191fea0d17617da01fb6e57eca827d0006fe34f07eaa54a53"} Mar 21 08:56:05 crc kubenswrapper[4696]: I0321 08:56:05.367256 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568056-sgpwk" Mar 21 08:56:05 crc kubenswrapper[4696]: I0321 08:56:05.454453 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xztc8\" (UniqueName: \"kubernetes.io/projected/f6936cdd-5be6-4162-a976-47d5df686290-kube-api-access-xztc8\") pod \"f6936cdd-5be6-4162-a976-47d5df686290\" (UID: \"f6936cdd-5be6-4162-a976-47d5df686290\") " Mar 21 08:56:05 crc kubenswrapper[4696]: I0321 08:56:05.463651 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6936cdd-5be6-4162-a976-47d5df686290-kube-api-access-xztc8" (OuterVolumeSpecName: "kube-api-access-xztc8") pod "f6936cdd-5be6-4162-a976-47d5df686290" (UID: "f6936cdd-5be6-4162-a976-47d5df686290"). InnerVolumeSpecName "kube-api-access-xztc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:56:05 crc kubenswrapper[4696]: I0321 08:56:05.557441 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xztc8\" (UniqueName: \"kubernetes.io/projected/f6936cdd-5be6-4162-a976-47d5df686290-kube-api-access-xztc8\") on node \"crc\" DevicePath \"\"" Mar 21 08:56:05 crc kubenswrapper[4696]: I0321 08:56:05.638809 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568056-sgpwk" event={"ID":"f6936cdd-5be6-4162-a976-47d5df686290","Type":"ContainerDied","Data":"8428f189f45f3e6479de883ed0e5eb1cc3bd18912e424b389d9a6892cfdf08ca"} Mar 21 08:56:05 crc kubenswrapper[4696]: I0321 08:56:05.638877 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8428f189f45f3e6479de883ed0e5eb1cc3bd18912e424b389d9a6892cfdf08ca" Mar 21 08:56:05 crc kubenswrapper[4696]: I0321 08:56:05.638964 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568056-sgpwk" Mar 21 08:56:05 crc kubenswrapper[4696]: I0321 08:56:05.710937 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568050-g4f92"] Mar 21 08:56:05 crc kubenswrapper[4696]: I0321 08:56:05.749202 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568050-g4f92"] Mar 21 08:56:06 crc kubenswrapper[4696]: I0321 08:56:06.546785 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e8563c8-4723-4e92-9972-3cddb4c529e7" path="/var/lib/kubelet/pods/7e8563c8-4723-4e92-9972-3cddb4c529e7/volumes" Mar 21 08:56:09 crc kubenswrapper[4696]: I0321 08:56:09.535190 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:56:09 crc kubenswrapper[4696]: E0321 08:56:09.535654 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:56:23 crc kubenswrapper[4696]: I0321 08:56:23.535437 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:56:23 crc kubenswrapper[4696]: E0321 08:56:23.536250 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:56:25 crc kubenswrapper[4696]: I0321 08:56:25.071257 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-7fs72"] Mar 21 08:56:25 crc kubenswrapper[4696]: I0321 08:56:25.094131 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-d153-account-create-update-728q9"] Mar 21 08:56:25 crc kubenswrapper[4696]: I0321 08:56:25.104147 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-b2hh7"] Mar 21 08:56:25 crc kubenswrapper[4696]: I0321 08:56:25.115158 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-nghsz"] Mar 21 08:56:25 crc kubenswrapper[4696]: I0321 08:56:25.125090 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-4f6e-account-create-update-wx2s5"] Mar 21 08:56:25 crc kubenswrapper[4696]: I0321 08:56:25.138131 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-552e-account-create-update-t4h4f"] Mar 21 08:56:25 crc kubenswrapper[4696]: I0321 08:56:25.151867 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-7fs72"] Mar 21 08:56:25 crc kubenswrapper[4696]: I0321 08:56:25.162139 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-nghsz"] Mar 21 08:56:25 crc kubenswrapper[4696]: I0321 08:56:25.174207 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-4f6e-account-create-update-wx2s5"] Mar 21 08:56:25 crc kubenswrapper[4696]: I0321 08:56:25.188343 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-552e-account-create-update-t4h4f"] Mar 21 08:56:25 crc kubenswrapper[4696]: I0321 08:56:25.198331 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-d153-account-create-update-728q9"] Mar 21 08:56:25 crc kubenswrapper[4696]: I0321 08:56:25.208345 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-b2hh7"] Mar 21 08:56:26 crc kubenswrapper[4696]: I0321 08:56:26.546484 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0227f04c-42f9-4a76-bd79-c1702200ca30" path="/var/lib/kubelet/pods/0227f04c-42f9-4a76-bd79-c1702200ca30/volumes" Mar 21 08:56:26 crc kubenswrapper[4696]: I0321 08:56:26.547078 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0936beb9-f71b-4313-b6a0-97b026f697a5" path="/var/lib/kubelet/pods/0936beb9-f71b-4313-b6a0-97b026f697a5/volumes" Mar 21 08:56:26 crc kubenswrapper[4696]: I0321 08:56:26.547624 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d5470b1-6bce-4dac-b1a7-f2ed50bcb713" path="/var/lib/kubelet/pods/0d5470b1-6bce-4dac-b1a7-f2ed50bcb713/volumes" Mar 21 08:56:26 crc kubenswrapper[4696]: I0321 08:56:26.548167 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c569fb8-edf7-4012-bd27-3eb30c568489" path="/var/lib/kubelet/pods/3c569fb8-edf7-4012-bd27-3eb30c568489/volumes" Mar 21 08:56:26 crc kubenswrapper[4696]: I0321 08:56:26.549236 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90f4f716-4da8-4ab0-8757-1c15480c73b5" path="/var/lib/kubelet/pods/90f4f716-4da8-4ab0-8757-1c15480c73b5/volumes" Mar 21 08:56:26 crc kubenswrapper[4696]: I0321 08:56:26.549736 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7067c05-68d2-413d-9580-af0eb2d31ef8" path="/var/lib/kubelet/pods/c7067c05-68d2-413d-9580-af0eb2d31ef8/volumes" Mar 21 08:56:38 crc kubenswrapper[4696]: I0321 08:56:38.535268 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:56:38 crc kubenswrapper[4696]: E0321 08:56:38.536218 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:56:48 crc kubenswrapper[4696]: I0321 08:56:48.032911 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-f359-account-create-update-2jqvh"] Mar 21 08:56:48 crc kubenswrapper[4696]: I0321 08:56:48.044684 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-mtbts"] Mar 21 08:56:48 crc kubenswrapper[4696]: I0321 08:56:48.053827 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-2ef2-account-create-update-9pxpg"] Mar 21 08:56:48 crc kubenswrapper[4696]: I0321 08:56:48.066715 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-mtbts"] Mar 21 08:56:48 crc kubenswrapper[4696]: I0321 08:56:48.077961 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-2ef2-account-create-update-9pxpg"] Mar 21 08:56:48 crc kubenswrapper[4696]: I0321 08:56:48.091456 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-f359-account-create-update-2jqvh"] Mar 21 08:56:48 crc kubenswrapper[4696]: I0321 08:56:48.546448 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17520989-db17-4ad3-8324-279698af93e3" path="/var/lib/kubelet/pods/17520989-db17-4ad3-8324-279698af93e3/volumes" Mar 21 08:56:48 crc kubenswrapper[4696]: I0321 08:56:48.547167 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5be1b1b8-ee1e-47cd-afbe-e1a96cee697a" path="/var/lib/kubelet/pods/5be1b1b8-ee1e-47cd-afbe-e1a96cee697a/volumes" Mar 21 08:56:48 crc kubenswrapper[4696]: I0321 08:56:48.547905 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec226d0f-1ee8-481b-85a5-280fe2d17810" path="/var/lib/kubelet/pods/ec226d0f-1ee8-481b-85a5-280fe2d17810/volumes" Mar 21 08:56:51 crc kubenswrapper[4696]: I0321 08:56:51.048969 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-52e3-account-create-update-5pmnh"] Mar 21 08:56:51 crc kubenswrapper[4696]: I0321 08:56:51.061613 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-4098-account-create-update-mn786"] Mar 21 08:56:51 crc kubenswrapper[4696]: I0321 08:56:51.071036 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-rff8n"] Mar 21 08:56:51 crc kubenswrapper[4696]: I0321 08:56:51.082180 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-jc8bj"] Mar 21 08:56:51 crc kubenswrapper[4696]: I0321 08:56:51.092108 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-4098-account-create-update-mn786"] Mar 21 08:56:51 crc kubenswrapper[4696]: I0321 08:56:51.103038 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-jc8bj"] Mar 21 08:56:51 crc kubenswrapper[4696]: I0321 08:56:51.117509 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-db-create-tjmmf"] Mar 21 08:56:51 crc kubenswrapper[4696]: I0321 08:56:51.132690 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-52e3-account-create-update-5pmnh"] Mar 21 08:56:51 crc kubenswrapper[4696]: I0321 08:56:51.143545 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-rff8n"] Mar 21 08:56:51 crc kubenswrapper[4696]: I0321 08:56:51.153957 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-db-create-tjmmf"] Mar 21 08:56:52 crc kubenswrapper[4696]: I0321 08:56:52.535350 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:56:52 crc kubenswrapper[4696]: E0321 08:56:52.535897 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:56:52 crc kubenswrapper[4696]: I0321 08:56:52.546024 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a6217ca-7143-41fd-ae0d-f96bde35029c" path="/var/lib/kubelet/pods/1a6217ca-7143-41fd-ae0d-f96bde35029c/volumes" Mar 21 08:56:52 crc kubenswrapper[4696]: I0321 08:56:52.546579 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80468ce8-5377-4423-9e04-b58d1d0910d7" path="/var/lib/kubelet/pods/80468ce8-5377-4423-9e04-b58d1d0910d7/volumes" Mar 21 08:56:52 crc kubenswrapper[4696]: I0321 08:56:52.547112 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bac528ca-04da-4a92-9b88-b9a3d226cb94" path="/var/lib/kubelet/pods/bac528ca-04da-4a92-9b88-b9a3d226cb94/volumes" Mar 21 08:56:52 crc kubenswrapper[4696]: I0321 08:56:52.547639 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf260a42-d042-4130-bc96-9e12dcb253c4" path="/var/lib/kubelet/pods/cf260a42-d042-4130-bc96-9e12dcb253c4/volumes" Mar 21 08:56:52 crc kubenswrapper[4696]: I0321 08:56:52.548618 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc7fc65b-70b4-46cd-84e5-16285d93f295" path="/var/lib/kubelet/pods/dc7fc65b-70b4-46cd-84e5-16285d93f295/volumes" Mar 21 08:56:59 crc kubenswrapper[4696]: I0321 08:56:59.033633 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-jqpz7"] Mar 21 08:56:59 crc kubenswrapper[4696]: I0321 08:56:59.037875 4696 scope.go:117] "RemoveContainer" containerID="95500aa101eb2803089e3ba8b625535edbe027bab63443bf2bfd2c0e85e980df" Mar 21 08:56:59 crc kubenswrapper[4696]: I0321 08:56:59.047580 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-jqpz7"] Mar 21 08:56:59 crc kubenswrapper[4696]: I0321 08:56:59.064294 4696 scope.go:117] "RemoveContainer" containerID="b7568e0277590b09dcd295ac32141b6adf5f79f58c74f5c1897d7e2d67b2a912" Mar 21 08:56:59 crc kubenswrapper[4696]: I0321 08:56:59.134791 4696 scope.go:117] "RemoveContainer" containerID="bd9b084d944210fce522b4d40b67d3fe59c3becdacac46cf7e660c5b96ede3d1" Mar 21 08:56:59 crc kubenswrapper[4696]: I0321 08:56:59.186514 4696 scope.go:117] "RemoveContainer" containerID="f4a25a7aa309127c8da648696529dc7f5675491cd67081147bf74654f558c98b" Mar 21 08:56:59 crc kubenswrapper[4696]: I0321 08:56:59.257465 4696 scope.go:117] "RemoveContainer" containerID="477c1fef1a311a6dee8cbdd3fb10286d5c2527cdd12e07df175ced7c294a0f79" Mar 21 08:56:59 crc kubenswrapper[4696]: I0321 08:56:59.294014 4696 scope.go:117] "RemoveContainer" containerID="f0b2b6e33fa875b269cbc9bceec63983ad8d84467cc4ec30c19f68f4d7c85243" Mar 21 08:56:59 crc kubenswrapper[4696]: I0321 08:56:59.338462 4696 scope.go:117] "RemoveContainer" containerID="090b80904642dc7de45b60aa7b56752e6360346514061cda1774d4693d550e96" Mar 21 08:56:59 crc kubenswrapper[4696]: I0321 08:56:59.357748 4696 scope.go:117] "RemoveContainer" containerID="0731f21de16f79a7267e66eb9659e169aec45ac89300ab26801aeb5d1d5d7aea" Mar 21 08:56:59 crc kubenswrapper[4696]: I0321 08:56:59.399020 4696 scope.go:117] "RemoveContainer" containerID="ff845693c915bd2420f675864f12930073f36bb1cfafce4dc20bad658632af06" Mar 21 08:56:59 crc kubenswrapper[4696]: I0321 08:56:59.417219 4696 scope.go:117] "RemoveContainer" containerID="7aaf2ff35fada122fcd81803dbb0c9200193736cb8409fff9c307db1291cbdf4" Mar 21 08:56:59 crc kubenswrapper[4696]: I0321 08:56:59.436273 4696 scope.go:117] "RemoveContainer" containerID="440899af6e87a066f77adc8aeb0aab34cd3a0ed754e6b67b3a2c4b827aeee757" Mar 21 08:56:59 crc kubenswrapper[4696]: I0321 08:56:59.459348 4696 scope.go:117] "RemoveContainer" containerID="2fbc10bfc4d089694e602949ff8f54651e8ede44ee5070f6939028e1a47a3406" Mar 21 08:56:59 crc kubenswrapper[4696]: I0321 08:56:59.481435 4696 scope.go:117] "RemoveContainer" containerID="165c8871ae76943c82c5256046d6b9992394e044b7f56566c6507b0d68eff7ac" Mar 21 08:56:59 crc kubenswrapper[4696]: I0321 08:56:59.506935 4696 scope.go:117] "RemoveContainer" containerID="1c41dad6f5faecac40e3d9cc1e1ff2d96ab5786733f4983487418a99056fe5f5" Mar 21 08:56:59 crc kubenswrapper[4696]: I0321 08:56:59.527080 4696 scope.go:117] "RemoveContainer" containerID="c2dd58cbddac3e8aa6f911a814016b9eb28687b142952a91d917b2d3bf923f6d" Mar 21 08:57:00 crc kubenswrapper[4696]: I0321 08:57:00.545606 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7359c71c-56ba-4c28-a126-bc95dea18dfd" path="/var/lib/kubelet/pods/7359c71c-56ba-4c28-a126-bc95dea18dfd/volumes" Mar 21 08:57:01 crc kubenswrapper[4696]: I0321 08:57:01.028113 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-trn7c"] Mar 21 08:57:01 crc kubenswrapper[4696]: I0321 08:57:01.036353 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-trn7c"] Mar 21 08:57:02 crc kubenswrapper[4696]: I0321 08:57:02.037686 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-d2jgl"] Mar 21 08:57:02 crc kubenswrapper[4696]: I0321 08:57:02.059740 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-d2jgl"] Mar 21 08:57:02 crc kubenswrapper[4696]: I0321 08:57:02.546265 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78946b93-1c47-42b0-98b4-7a9454ea8dbc" path="/var/lib/kubelet/pods/78946b93-1c47-42b0-98b4-7a9454ea8dbc/volumes" Mar 21 08:57:02 crc kubenswrapper[4696]: I0321 08:57:02.547024 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92718cf1-7a8f-42cd-aadc-df800f2499d4" path="/var/lib/kubelet/pods/92718cf1-7a8f-42cd-aadc-df800f2499d4/volumes" Mar 21 08:57:03 crc kubenswrapper[4696]: I0321 08:57:03.535438 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:57:03 crc kubenswrapper[4696]: E0321 08:57:03.536197 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:57:15 crc kubenswrapper[4696]: I0321 08:57:15.534936 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:57:15 crc kubenswrapper[4696]: E0321 08:57:15.537025 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:57:20 crc kubenswrapper[4696]: I0321 08:57:20.233341 4696 generic.go:334] "Generic (PLEG): container finished" podID="c756d18c-c925-44bd-911f-0b82d961c66e" containerID="37990705b78f5cf85a469b6c4ccc626bee25e36ae7fda2c36575e1283f2f1bf7" exitCode=0 Mar 21 08:57:20 crc kubenswrapper[4696]: I0321 08:57:20.233463 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zncdp" event={"ID":"c756d18c-c925-44bd-911f-0b82d961c66e","Type":"ContainerDied","Data":"37990705b78f5cf85a469b6c4ccc626bee25e36ae7fda2c36575e1283f2f1bf7"} Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.158634 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zncdp" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.258659 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zncdp" event={"ID":"c756d18c-c925-44bd-911f-0b82d961c66e","Type":"ContainerDied","Data":"7de3e383f017f3451359a7ea99d48ef414403843f1833f29bfc0f7dac18856c1"} Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.258696 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7de3e383f017f3451359a7ea99d48ef414403843f1833f29bfc0f7dac18856c1" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.258747 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zncdp" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.261333 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8hxt\" (UniqueName: \"kubernetes.io/projected/c756d18c-c925-44bd-911f-0b82d961c66e-kube-api-access-n8hxt\") pod \"c756d18c-c925-44bd-911f-0b82d961c66e\" (UID: \"c756d18c-c925-44bd-911f-0b82d961c66e\") " Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.261800 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c756d18c-c925-44bd-911f-0b82d961c66e-ssh-key-openstack-edpm-ipam\") pod \"c756d18c-c925-44bd-911f-0b82d961c66e\" (UID: \"c756d18c-c925-44bd-911f-0b82d961c66e\") " Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.262034 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c756d18c-c925-44bd-911f-0b82d961c66e-inventory\") pod \"c756d18c-c925-44bd-911f-0b82d961c66e\" (UID: \"c756d18c-c925-44bd-911f-0b82d961c66e\") " Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.267020 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c756d18c-c925-44bd-911f-0b82d961c66e-kube-api-access-n8hxt" (OuterVolumeSpecName: "kube-api-access-n8hxt") pod "c756d18c-c925-44bd-911f-0b82d961c66e" (UID: "c756d18c-c925-44bd-911f-0b82d961c66e"). InnerVolumeSpecName "kube-api-access-n8hxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.303028 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c756d18c-c925-44bd-911f-0b82d961c66e-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c756d18c-c925-44bd-911f-0b82d961c66e" (UID: "c756d18c-c925-44bd-911f-0b82d961c66e"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.334126 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c756d18c-c925-44bd-911f-0b82d961c66e-inventory" (OuterVolumeSpecName: "inventory") pod "c756d18c-c925-44bd-911f-0b82d961c66e" (UID: "c756d18c-c925-44bd-911f-0b82d961c66e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.347865 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv"] Mar 21 08:57:22 crc kubenswrapper[4696]: E0321 08:57:22.348299 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c756d18c-c925-44bd-911f-0b82d961c66e" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.348315 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="c756d18c-c925-44bd-911f-0b82d961c66e" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 21 08:57:22 crc kubenswrapper[4696]: E0321 08:57:22.348352 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6936cdd-5be6-4162-a976-47d5df686290" containerName="oc" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.348358 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6936cdd-5be6-4162-a976-47d5df686290" containerName="oc" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.348560 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="c756d18c-c925-44bd-911f-0b82d961c66e" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.348592 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6936cdd-5be6-4162-a976-47d5df686290" containerName="oc" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.349387 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.364548 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c756d18c-c925-44bd-911f-0b82d961c66e-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.364579 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c756d18c-c925-44bd-911f-0b82d961c66e-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.364589 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8hxt\" (UniqueName: \"kubernetes.io/projected/c756d18c-c925-44bd-911f-0b82d961c66e-kube-api-access-n8hxt\") on node \"crc\" DevicePath \"\"" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.365437 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv"] Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.466684 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/719c50b9-02ed-4138-8ce6-624b718e5d00-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv\" (UID: \"719c50b9-02ed-4138-8ce6-624b718e5d00\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.466864 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nslz5\" (UniqueName: \"kubernetes.io/projected/719c50b9-02ed-4138-8ce6-624b718e5d00-kube-api-access-nslz5\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv\" (UID: \"719c50b9-02ed-4138-8ce6-624b718e5d00\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.466916 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/719c50b9-02ed-4138-8ce6-624b718e5d00-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv\" (UID: \"719c50b9-02ed-4138-8ce6-624b718e5d00\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.568634 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/719c50b9-02ed-4138-8ce6-624b718e5d00-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv\" (UID: \"719c50b9-02ed-4138-8ce6-624b718e5d00\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.568760 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nslz5\" (UniqueName: \"kubernetes.io/projected/719c50b9-02ed-4138-8ce6-624b718e5d00-kube-api-access-nslz5\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv\" (UID: \"719c50b9-02ed-4138-8ce6-624b718e5d00\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.568804 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/719c50b9-02ed-4138-8ce6-624b718e5d00-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv\" (UID: \"719c50b9-02ed-4138-8ce6-624b718e5d00\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.572466 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/719c50b9-02ed-4138-8ce6-624b718e5d00-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv\" (UID: \"719c50b9-02ed-4138-8ce6-624b718e5d00\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.573430 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/719c50b9-02ed-4138-8ce6-624b718e5d00-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv\" (UID: \"719c50b9-02ed-4138-8ce6-624b718e5d00\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.588725 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nslz5\" (UniqueName: \"kubernetes.io/projected/719c50b9-02ed-4138-8ce6-624b718e5d00-kube-api-access-nslz5\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv\" (UID: \"719c50b9-02ed-4138-8ce6-624b718e5d00\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv" Mar 21 08:57:22 crc kubenswrapper[4696]: I0321 08:57:22.737803 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv" Mar 21 08:57:23 crc kubenswrapper[4696]: I0321 08:57:23.593309 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv"] Mar 21 08:57:23 crc kubenswrapper[4696]: I0321 08:57:23.602238 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 08:57:24 crc kubenswrapper[4696]: I0321 08:57:24.279507 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv" event={"ID":"719c50b9-02ed-4138-8ce6-624b718e5d00","Type":"ContainerStarted","Data":"1223a99635ba79bdbf12681ff3637ec82020e6401d11d6095f9e9e76bae75ae9"} Mar 21 08:57:25 crc kubenswrapper[4696]: I0321 08:57:25.290325 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv" event={"ID":"719c50b9-02ed-4138-8ce6-624b718e5d00","Type":"ContainerStarted","Data":"1e280df2b7d9420ed01c1808a0e34d22cab673aea87c56a5d97d266e7b86d5d1"} Mar 21 08:57:25 crc kubenswrapper[4696]: I0321 08:57:25.310848 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv" podStartSLOduration=2.860302625 podStartE2EDuration="3.310807132s" podCreationTimestamp="2026-03-21 08:57:22 +0000 UTC" firstStartedPulling="2026-03-21 08:57:23.601859285 +0000 UTC m=+1777.722739998" lastFinishedPulling="2026-03-21 08:57:24.052363792 +0000 UTC m=+1778.173244505" observedRunningTime="2026-03-21 08:57:25.307194703 +0000 UTC m=+1779.428075416" watchObservedRunningTime="2026-03-21 08:57:25.310807132 +0000 UTC m=+1779.431687845" Mar 21 08:57:27 crc kubenswrapper[4696]: I0321 08:57:27.535873 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:57:27 crc kubenswrapper[4696]: E0321 08:57:27.536425 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:57:36 crc kubenswrapper[4696]: I0321 08:57:36.059344 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-m25wk"] Mar 21 08:57:36 crc kubenswrapper[4696]: I0321 08:57:36.072491 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-m25wk"] Mar 21 08:57:36 crc kubenswrapper[4696]: I0321 08:57:36.546749 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4ac0955-0d2a-4081-b82b-32dd288de552" path="/var/lib/kubelet/pods/c4ac0955-0d2a-4081-b82b-32dd288de552/volumes" Mar 21 08:57:37 crc kubenswrapper[4696]: I0321 08:57:37.250165 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ksbn7"] Mar 21 08:57:37 crc kubenswrapper[4696]: I0321 08:57:37.252608 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ksbn7" Mar 21 08:57:37 crc kubenswrapper[4696]: I0321 08:57:37.261534 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ksbn7"] Mar 21 08:57:37 crc kubenswrapper[4696]: I0321 08:57:37.416154 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhw49\" (UniqueName: \"kubernetes.io/projected/13321b01-6e04-448f-bb8c-036f033d5dba-kube-api-access-hhw49\") pod \"community-operators-ksbn7\" (UID: \"13321b01-6e04-448f-bb8c-036f033d5dba\") " pod="openshift-marketplace/community-operators-ksbn7" Mar 21 08:57:37 crc kubenswrapper[4696]: I0321 08:57:37.416243 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13321b01-6e04-448f-bb8c-036f033d5dba-catalog-content\") pod \"community-operators-ksbn7\" (UID: \"13321b01-6e04-448f-bb8c-036f033d5dba\") " pod="openshift-marketplace/community-operators-ksbn7" Mar 21 08:57:37 crc kubenswrapper[4696]: I0321 08:57:37.416338 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13321b01-6e04-448f-bb8c-036f033d5dba-utilities\") pod \"community-operators-ksbn7\" (UID: \"13321b01-6e04-448f-bb8c-036f033d5dba\") " pod="openshift-marketplace/community-operators-ksbn7" Mar 21 08:57:37 crc kubenswrapper[4696]: I0321 08:57:37.518137 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhw49\" (UniqueName: \"kubernetes.io/projected/13321b01-6e04-448f-bb8c-036f033d5dba-kube-api-access-hhw49\") pod \"community-operators-ksbn7\" (UID: \"13321b01-6e04-448f-bb8c-036f033d5dba\") " pod="openshift-marketplace/community-operators-ksbn7" Mar 21 08:57:37 crc kubenswrapper[4696]: I0321 08:57:37.518532 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13321b01-6e04-448f-bb8c-036f033d5dba-catalog-content\") pod \"community-operators-ksbn7\" (UID: \"13321b01-6e04-448f-bb8c-036f033d5dba\") " pod="openshift-marketplace/community-operators-ksbn7" Mar 21 08:57:37 crc kubenswrapper[4696]: I0321 08:57:37.518780 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13321b01-6e04-448f-bb8c-036f033d5dba-utilities\") pod \"community-operators-ksbn7\" (UID: \"13321b01-6e04-448f-bb8c-036f033d5dba\") " pod="openshift-marketplace/community-operators-ksbn7" Mar 21 08:57:37 crc kubenswrapper[4696]: I0321 08:57:37.518976 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13321b01-6e04-448f-bb8c-036f033d5dba-catalog-content\") pod \"community-operators-ksbn7\" (UID: \"13321b01-6e04-448f-bb8c-036f033d5dba\") " pod="openshift-marketplace/community-operators-ksbn7" Mar 21 08:57:37 crc kubenswrapper[4696]: I0321 08:57:37.519222 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13321b01-6e04-448f-bb8c-036f033d5dba-utilities\") pod \"community-operators-ksbn7\" (UID: \"13321b01-6e04-448f-bb8c-036f033d5dba\") " pod="openshift-marketplace/community-operators-ksbn7" Mar 21 08:57:37 crc kubenswrapper[4696]: I0321 08:57:37.549305 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhw49\" (UniqueName: \"kubernetes.io/projected/13321b01-6e04-448f-bb8c-036f033d5dba-kube-api-access-hhw49\") pod \"community-operators-ksbn7\" (UID: \"13321b01-6e04-448f-bb8c-036f033d5dba\") " pod="openshift-marketplace/community-operators-ksbn7" Mar 21 08:57:37 crc kubenswrapper[4696]: I0321 08:57:37.573151 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ksbn7" Mar 21 08:57:38 crc kubenswrapper[4696]: I0321 08:57:38.151429 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ksbn7"] Mar 21 08:57:38 crc kubenswrapper[4696]: I0321 08:57:38.424594 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ksbn7" event={"ID":"13321b01-6e04-448f-bb8c-036f033d5dba","Type":"ContainerStarted","Data":"6aed6235da32404a3b035f058de86235f49af8c36ceff1e3e4813c7632f3514d"} Mar 21 08:57:39 crc kubenswrapper[4696]: I0321 08:57:39.447300 4696 generic.go:334] "Generic (PLEG): container finished" podID="13321b01-6e04-448f-bb8c-036f033d5dba" containerID="fb6b3ec27889b3029a8baa640d197d379990a1d281248b6f92814d918162167b" exitCode=0 Mar 21 08:57:39 crc kubenswrapper[4696]: I0321 08:57:39.447483 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ksbn7" event={"ID":"13321b01-6e04-448f-bb8c-036f033d5dba","Type":"ContainerDied","Data":"fb6b3ec27889b3029a8baa640d197d379990a1d281248b6f92814d918162167b"} Mar 21 08:57:40 crc kubenswrapper[4696]: I0321 08:57:40.458494 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ksbn7" event={"ID":"13321b01-6e04-448f-bb8c-036f033d5dba","Type":"ContainerStarted","Data":"b7dae7638be09a938dc3ad8df3675857234702ff7367ac9df086ff48263b64fb"} Mar 21 08:57:42 crc kubenswrapper[4696]: I0321 08:57:42.477451 4696 generic.go:334] "Generic (PLEG): container finished" podID="13321b01-6e04-448f-bb8c-036f033d5dba" containerID="b7dae7638be09a938dc3ad8df3675857234702ff7367ac9df086ff48263b64fb" exitCode=0 Mar 21 08:57:42 crc kubenswrapper[4696]: I0321 08:57:42.477531 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ksbn7" event={"ID":"13321b01-6e04-448f-bb8c-036f033d5dba","Type":"ContainerDied","Data":"b7dae7638be09a938dc3ad8df3675857234702ff7367ac9df086ff48263b64fb"} Mar 21 08:57:42 crc kubenswrapper[4696]: I0321 08:57:42.535218 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:57:42 crc kubenswrapper[4696]: E0321 08:57:42.535460 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:57:43 crc kubenswrapper[4696]: I0321 08:57:43.492493 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ksbn7" event={"ID":"13321b01-6e04-448f-bb8c-036f033d5dba","Type":"ContainerStarted","Data":"862dfc60199fb3e0db1dd026b4321564228dd8fd00639d45af17f8093bb76ac1"} Mar 21 08:57:43 crc kubenswrapper[4696]: I0321 08:57:43.519510 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ksbn7" podStartSLOduration=3.024986918 podStartE2EDuration="6.519479951s" podCreationTimestamp="2026-03-21 08:57:37 +0000 UTC" firstStartedPulling="2026-03-21 08:57:39.450264598 +0000 UTC m=+1793.571145311" lastFinishedPulling="2026-03-21 08:57:42.944757631 +0000 UTC m=+1797.065638344" observedRunningTime="2026-03-21 08:57:43.510502636 +0000 UTC m=+1797.631383339" watchObservedRunningTime="2026-03-21 08:57:43.519479951 +0000 UTC m=+1797.640360664" Mar 21 08:57:46 crc kubenswrapper[4696]: I0321 08:57:46.039291 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-nqvgp"] Mar 21 08:57:46 crc kubenswrapper[4696]: I0321 08:57:46.052451 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-nqvgp"] Mar 21 08:57:46 crc kubenswrapper[4696]: I0321 08:57:46.549020 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80364601-ba09-4192-a176-e04ad8d45506" path="/var/lib/kubelet/pods/80364601-ba09-4192-a176-e04ad8d45506/volumes" Mar 21 08:57:47 crc kubenswrapper[4696]: I0321 08:57:47.035331 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-cwvf9"] Mar 21 08:57:47 crc kubenswrapper[4696]: I0321 08:57:47.048826 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-cwvf9"] Mar 21 08:57:47 crc kubenswrapper[4696]: I0321 08:57:47.573648 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ksbn7" Mar 21 08:57:47 crc kubenswrapper[4696]: I0321 08:57:47.573697 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ksbn7" Mar 21 08:57:47 crc kubenswrapper[4696]: I0321 08:57:47.641512 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ksbn7" Mar 21 08:57:48 crc kubenswrapper[4696]: I0321 08:57:48.035499 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-rzcz6"] Mar 21 08:57:48 crc kubenswrapper[4696]: I0321 08:57:48.049122 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-rzcz6"] Mar 21 08:57:48 crc kubenswrapper[4696]: I0321 08:57:48.556283 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4381bdfb-820a-42cc-9a0f-0ca3e6f34449" path="/var/lib/kubelet/pods/4381bdfb-820a-42cc-9a0f-0ca3e6f34449/volumes" Mar 21 08:57:48 crc kubenswrapper[4696]: I0321 08:57:48.557370 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8556431-841c-4372-8c93-fc3649d9b0f0" path="/var/lib/kubelet/pods/b8556431-841c-4372-8c93-fc3649d9b0f0/volumes" Mar 21 08:57:48 crc kubenswrapper[4696]: I0321 08:57:48.594335 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ksbn7" Mar 21 08:57:51 crc kubenswrapper[4696]: I0321 08:57:51.234768 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ksbn7"] Mar 21 08:57:51 crc kubenswrapper[4696]: I0321 08:57:51.235285 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ksbn7" podUID="13321b01-6e04-448f-bb8c-036f033d5dba" containerName="registry-server" containerID="cri-o://862dfc60199fb3e0db1dd026b4321564228dd8fd00639d45af17f8093bb76ac1" gracePeriod=2 Mar 21 08:57:51 crc kubenswrapper[4696]: I0321 08:57:51.577510 4696 generic.go:334] "Generic (PLEG): container finished" podID="13321b01-6e04-448f-bb8c-036f033d5dba" containerID="862dfc60199fb3e0db1dd026b4321564228dd8fd00639d45af17f8093bb76ac1" exitCode=0 Mar 21 08:57:51 crc kubenswrapper[4696]: I0321 08:57:51.577577 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ksbn7" event={"ID":"13321b01-6e04-448f-bb8c-036f033d5dba","Type":"ContainerDied","Data":"862dfc60199fb3e0db1dd026b4321564228dd8fd00639d45af17f8093bb76ac1"} Mar 21 08:57:52 crc kubenswrapper[4696]: I0321 08:57:52.275859 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ksbn7" Mar 21 08:57:52 crc kubenswrapper[4696]: I0321 08:57:52.414108 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhw49\" (UniqueName: \"kubernetes.io/projected/13321b01-6e04-448f-bb8c-036f033d5dba-kube-api-access-hhw49\") pod \"13321b01-6e04-448f-bb8c-036f033d5dba\" (UID: \"13321b01-6e04-448f-bb8c-036f033d5dba\") " Mar 21 08:57:52 crc kubenswrapper[4696]: I0321 08:57:52.414408 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13321b01-6e04-448f-bb8c-036f033d5dba-utilities\") pod \"13321b01-6e04-448f-bb8c-036f033d5dba\" (UID: \"13321b01-6e04-448f-bb8c-036f033d5dba\") " Mar 21 08:57:52 crc kubenswrapper[4696]: I0321 08:57:52.414509 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13321b01-6e04-448f-bb8c-036f033d5dba-catalog-content\") pod \"13321b01-6e04-448f-bb8c-036f033d5dba\" (UID: \"13321b01-6e04-448f-bb8c-036f033d5dba\") " Mar 21 08:57:52 crc kubenswrapper[4696]: I0321 08:57:52.414911 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13321b01-6e04-448f-bb8c-036f033d5dba-utilities" (OuterVolumeSpecName: "utilities") pod "13321b01-6e04-448f-bb8c-036f033d5dba" (UID: "13321b01-6e04-448f-bb8c-036f033d5dba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:57:52 crc kubenswrapper[4696]: I0321 08:57:52.415141 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13321b01-6e04-448f-bb8c-036f033d5dba-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 08:57:52 crc kubenswrapper[4696]: I0321 08:57:52.420621 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13321b01-6e04-448f-bb8c-036f033d5dba-kube-api-access-hhw49" (OuterVolumeSpecName: "kube-api-access-hhw49") pod "13321b01-6e04-448f-bb8c-036f033d5dba" (UID: "13321b01-6e04-448f-bb8c-036f033d5dba"). InnerVolumeSpecName "kube-api-access-hhw49". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:57:52 crc kubenswrapper[4696]: I0321 08:57:52.472235 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13321b01-6e04-448f-bb8c-036f033d5dba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "13321b01-6e04-448f-bb8c-036f033d5dba" (UID: "13321b01-6e04-448f-bb8c-036f033d5dba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 08:57:52 crc kubenswrapper[4696]: I0321 08:57:52.517794 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13321b01-6e04-448f-bb8c-036f033d5dba-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 08:57:52 crc kubenswrapper[4696]: I0321 08:57:52.517856 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhw49\" (UniqueName: \"kubernetes.io/projected/13321b01-6e04-448f-bb8c-036f033d5dba-kube-api-access-hhw49\") on node \"crc\" DevicePath \"\"" Mar 21 08:57:52 crc kubenswrapper[4696]: I0321 08:57:52.589861 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ksbn7" event={"ID":"13321b01-6e04-448f-bb8c-036f033d5dba","Type":"ContainerDied","Data":"6aed6235da32404a3b035f058de86235f49af8c36ceff1e3e4813c7632f3514d"} Mar 21 08:57:52 crc kubenswrapper[4696]: I0321 08:57:52.589904 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ksbn7" Mar 21 08:57:52 crc kubenswrapper[4696]: I0321 08:57:52.589931 4696 scope.go:117] "RemoveContainer" containerID="862dfc60199fb3e0db1dd026b4321564228dd8fd00639d45af17f8093bb76ac1" Mar 21 08:57:52 crc kubenswrapper[4696]: I0321 08:57:52.616006 4696 scope.go:117] "RemoveContainer" containerID="b7dae7638be09a938dc3ad8df3675857234702ff7367ac9df086ff48263b64fb" Mar 21 08:57:52 crc kubenswrapper[4696]: I0321 08:57:52.621903 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ksbn7"] Mar 21 08:57:52 crc kubenswrapper[4696]: I0321 08:57:52.629166 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ksbn7"] Mar 21 08:57:52 crc kubenswrapper[4696]: I0321 08:57:52.644804 4696 scope.go:117] "RemoveContainer" containerID="fb6b3ec27889b3029a8baa640d197d379990a1d281248b6f92814d918162167b" Mar 21 08:57:54 crc kubenswrapper[4696]: I0321 08:57:54.534918 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:57:54 crc kubenswrapper[4696]: E0321 08:57:54.535240 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:57:54 crc kubenswrapper[4696]: I0321 08:57:54.545707 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13321b01-6e04-448f-bb8c-036f033d5dba" path="/var/lib/kubelet/pods/13321b01-6e04-448f-bb8c-036f033d5dba/volumes" Mar 21 08:57:59 crc kubenswrapper[4696]: I0321 08:57:59.824015 4696 scope.go:117] "RemoveContainer" containerID="76e69cc985d933b3ea0095ce375e81e97f585398126f36881f496d0ff4681010" Mar 21 08:57:59 crc kubenswrapper[4696]: I0321 08:57:59.850046 4696 scope.go:117] "RemoveContainer" containerID="39e6b99fa4de32a798f582ebba0755776e9fe76d1fef544f75ba6269ef3aa353" Mar 21 08:57:59 crc kubenswrapper[4696]: I0321 08:57:59.920049 4696 scope.go:117] "RemoveContainer" containerID="23bd157c07ed6f47aae360143f80c9569b30775dd932dfdee1c2a28699730de9" Mar 21 08:58:00 crc kubenswrapper[4696]: I0321 08:58:00.007072 4696 scope.go:117] "RemoveContainer" containerID="c43fd1da93fdfad7e2a19752084be7a45719936890db99fe36fff7b8af9b358e" Mar 21 08:58:00 crc kubenswrapper[4696]: I0321 08:58:00.034881 4696 scope.go:117] "RemoveContainer" containerID="47939aa8ea3147e4b4b4c629c0914bbb4bd10f5faf5fc4b3728caa3b9d4324cb" Mar 21 08:58:00 crc kubenswrapper[4696]: I0321 08:58:00.095397 4696 scope.go:117] "RemoveContainer" containerID="00da315452c455fa030a1e82cea1365333e9da4c1223f6f94970f47b1284776e" Mar 21 08:58:00 crc kubenswrapper[4696]: I0321 08:58:00.144452 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568058-9z5qs"] Mar 21 08:58:00 crc kubenswrapper[4696]: E0321 08:58:00.144880 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13321b01-6e04-448f-bb8c-036f033d5dba" containerName="extract-content" Mar 21 08:58:00 crc kubenswrapper[4696]: I0321 08:58:00.144895 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="13321b01-6e04-448f-bb8c-036f033d5dba" containerName="extract-content" Mar 21 08:58:00 crc kubenswrapper[4696]: E0321 08:58:00.144908 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13321b01-6e04-448f-bb8c-036f033d5dba" containerName="registry-server" Mar 21 08:58:00 crc kubenswrapper[4696]: I0321 08:58:00.144915 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="13321b01-6e04-448f-bb8c-036f033d5dba" containerName="registry-server" Mar 21 08:58:00 crc kubenswrapper[4696]: E0321 08:58:00.144950 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13321b01-6e04-448f-bb8c-036f033d5dba" containerName="extract-utilities" Mar 21 08:58:00 crc kubenswrapper[4696]: I0321 08:58:00.144959 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="13321b01-6e04-448f-bb8c-036f033d5dba" containerName="extract-utilities" Mar 21 08:58:00 crc kubenswrapper[4696]: I0321 08:58:00.145155 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="13321b01-6e04-448f-bb8c-036f033d5dba" containerName="registry-server" Mar 21 08:58:00 crc kubenswrapper[4696]: I0321 08:58:00.145882 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568058-9z5qs" Mar 21 08:58:00 crc kubenswrapper[4696]: I0321 08:58:00.151283 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 08:58:00 crc kubenswrapper[4696]: I0321 08:58:00.151511 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 08:58:00 crc kubenswrapper[4696]: I0321 08:58:00.151651 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 08:58:00 crc kubenswrapper[4696]: I0321 08:58:00.155185 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568058-9z5qs"] Mar 21 08:58:00 crc kubenswrapper[4696]: I0321 08:58:00.159372 4696 scope.go:117] "RemoveContainer" containerID="f07398703df0547d265b0236bc994a5f3e2b860e2c22130759ce02432acf9cdf" Mar 21 08:58:00 crc kubenswrapper[4696]: I0321 08:58:00.187865 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhjhm\" (UniqueName: \"kubernetes.io/projected/6da51675-21b7-4468-b86f-28e19ec86020-kube-api-access-dhjhm\") pod \"auto-csr-approver-29568058-9z5qs\" (UID: \"6da51675-21b7-4468-b86f-28e19ec86020\") " pod="openshift-infra/auto-csr-approver-29568058-9z5qs" Mar 21 08:58:00 crc kubenswrapper[4696]: I0321 08:58:00.290505 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhjhm\" (UniqueName: \"kubernetes.io/projected/6da51675-21b7-4468-b86f-28e19ec86020-kube-api-access-dhjhm\") pod \"auto-csr-approver-29568058-9z5qs\" (UID: \"6da51675-21b7-4468-b86f-28e19ec86020\") " pod="openshift-infra/auto-csr-approver-29568058-9z5qs" Mar 21 08:58:00 crc kubenswrapper[4696]: I0321 08:58:00.307180 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhjhm\" (UniqueName: \"kubernetes.io/projected/6da51675-21b7-4468-b86f-28e19ec86020-kube-api-access-dhjhm\") pod \"auto-csr-approver-29568058-9z5qs\" (UID: \"6da51675-21b7-4468-b86f-28e19ec86020\") " pod="openshift-infra/auto-csr-approver-29568058-9z5qs" Mar 21 08:58:00 crc kubenswrapper[4696]: I0321 08:58:00.504067 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568058-9z5qs" Mar 21 08:58:01 crc kubenswrapper[4696]: I0321 08:58:01.229034 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568058-9z5qs"] Mar 21 08:58:01 crc kubenswrapper[4696]: I0321 08:58:01.694543 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568058-9z5qs" event={"ID":"6da51675-21b7-4468-b86f-28e19ec86020","Type":"ContainerStarted","Data":"1c2ff16c36c04acda609a15a0a43ce06a112e6c71bccc8a4a54770abefad5913"} Mar 21 08:58:02 crc kubenswrapper[4696]: I0321 08:58:02.711115 4696 generic.go:334] "Generic (PLEG): container finished" podID="6da51675-21b7-4468-b86f-28e19ec86020" containerID="fcc17446e6d74173152ee59e6caa9c2656bf271716767e63d5662f0be42c52a3" exitCode=0 Mar 21 08:58:02 crc kubenswrapper[4696]: I0321 08:58:02.711303 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568058-9z5qs" event={"ID":"6da51675-21b7-4468-b86f-28e19ec86020","Type":"ContainerDied","Data":"fcc17446e6d74173152ee59e6caa9c2656bf271716767e63d5662f0be42c52a3"} Mar 21 08:58:04 crc kubenswrapper[4696]: I0321 08:58:04.563161 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568058-9z5qs" Mar 21 08:58:04 crc kubenswrapper[4696]: I0321 08:58:04.684191 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhjhm\" (UniqueName: \"kubernetes.io/projected/6da51675-21b7-4468-b86f-28e19ec86020-kube-api-access-dhjhm\") pod \"6da51675-21b7-4468-b86f-28e19ec86020\" (UID: \"6da51675-21b7-4468-b86f-28e19ec86020\") " Mar 21 08:58:04 crc kubenswrapper[4696]: I0321 08:58:04.701011 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6da51675-21b7-4468-b86f-28e19ec86020-kube-api-access-dhjhm" (OuterVolumeSpecName: "kube-api-access-dhjhm") pod "6da51675-21b7-4468-b86f-28e19ec86020" (UID: "6da51675-21b7-4468-b86f-28e19ec86020"). InnerVolumeSpecName "kube-api-access-dhjhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:58:04 crc kubenswrapper[4696]: I0321 08:58:04.728950 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568058-9z5qs" event={"ID":"6da51675-21b7-4468-b86f-28e19ec86020","Type":"ContainerDied","Data":"1c2ff16c36c04acda609a15a0a43ce06a112e6c71bccc8a4a54770abefad5913"} Mar 21 08:58:04 crc kubenswrapper[4696]: I0321 08:58:04.729011 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c2ff16c36c04acda609a15a0a43ce06a112e6c71bccc8a4a54770abefad5913" Mar 21 08:58:04 crc kubenswrapper[4696]: I0321 08:58:04.729022 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568058-9z5qs" Mar 21 08:58:04 crc kubenswrapper[4696]: I0321 08:58:04.787118 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhjhm\" (UniqueName: \"kubernetes.io/projected/6da51675-21b7-4468-b86f-28e19ec86020-kube-api-access-dhjhm\") on node \"crc\" DevicePath \"\"" Mar 21 08:58:05 crc kubenswrapper[4696]: I0321 08:58:05.641031 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568052-n4r22"] Mar 21 08:58:05 crc kubenswrapper[4696]: I0321 08:58:05.652315 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568052-n4r22"] Mar 21 08:58:06 crc kubenswrapper[4696]: I0321 08:58:06.545273 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cdcde53-8a3a-449a-a525-fccb10e196aa" path="/var/lib/kubelet/pods/0cdcde53-8a3a-449a-a525-fccb10e196aa/volumes" Mar 21 08:58:08 crc kubenswrapper[4696]: I0321 08:58:08.027854 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-n5tdf"] Mar 21 08:58:08 crc kubenswrapper[4696]: I0321 08:58:08.040211 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-n5tdf"] Mar 21 08:58:08 crc kubenswrapper[4696]: I0321 08:58:08.547949 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a70d1c8a-2db3-4ade-af61-fc8225262c18" path="/var/lib/kubelet/pods/a70d1c8a-2db3-4ade-af61-fc8225262c18/volumes" Mar 21 08:58:09 crc kubenswrapper[4696]: I0321 08:58:09.535233 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:58:09 crc kubenswrapper[4696]: E0321 08:58:09.535704 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:58:22 crc kubenswrapper[4696]: I0321 08:58:22.534907 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:58:22 crc kubenswrapper[4696]: E0321 08:58:22.535670 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:58:33 crc kubenswrapper[4696]: I0321 08:58:33.034528 4696 generic.go:334] "Generic (PLEG): container finished" podID="719c50b9-02ed-4138-8ce6-624b718e5d00" containerID="1e280df2b7d9420ed01c1808a0e34d22cab673aea87c56a5d97d266e7b86d5d1" exitCode=0 Mar 21 08:58:33 crc kubenswrapper[4696]: I0321 08:58:33.034621 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv" event={"ID":"719c50b9-02ed-4138-8ce6-624b718e5d00","Type":"ContainerDied","Data":"1e280df2b7d9420ed01c1808a0e34d22cab673aea87c56a5d97d266e7b86d5d1"} Mar 21 08:58:34 crc kubenswrapper[4696]: I0321 08:58:34.892519 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv" Mar 21 08:58:34 crc kubenswrapper[4696]: I0321 08:58:34.998800 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/719c50b9-02ed-4138-8ce6-624b718e5d00-ssh-key-openstack-edpm-ipam\") pod \"719c50b9-02ed-4138-8ce6-624b718e5d00\" (UID: \"719c50b9-02ed-4138-8ce6-624b718e5d00\") " Mar 21 08:58:34 crc kubenswrapper[4696]: I0321 08:58:34.999472 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nslz5\" (UniqueName: \"kubernetes.io/projected/719c50b9-02ed-4138-8ce6-624b718e5d00-kube-api-access-nslz5\") pod \"719c50b9-02ed-4138-8ce6-624b718e5d00\" (UID: \"719c50b9-02ed-4138-8ce6-624b718e5d00\") " Mar 21 08:58:34 crc kubenswrapper[4696]: I0321 08:58:34.999540 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/719c50b9-02ed-4138-8ce6-624b718e5d00-inventory\") pod \"719c50b9-02ed-4138-8ce6-624b718e5d00\" (UID: \"719c50b9-02ed-4138-8ce6-624b718e5d00\") " Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.018704 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/719c50b9-02ed-4138-8ce6-624b718e5d00-kube-api-access-nslz5" (OuterVolumeSpecName: "kube-api-access-nslz5") pod "719c50b9-02ed-4138-8ce6-624b718e5d00" (UID: "719c50b9-02ed-4138-8ce6-624b718e5d00"). InnerVolumeSpecName "kube-api-access-nslz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.028875 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/719c50b9-02ed-4138-8ce6-624b718e5d00-inventory" (OuterVolumeSpecName: "inventory") pod "719c50b9-02ed-4138-8ce6-624b718e5d00" (UID: "719c50b9-02ed-4138-8ce6-624b718e5d00"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.033958 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/719c50b9-02ed-4138-8ce6-624b718e5d00-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "719c50b9-02ed-4138-8ce6-624b718e5d00" (UID: "719c50b9-02ed-4138-8ce6-624b718e5d00"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.072922 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv" event={"ID":"719c50b9-02ed-4138-8ce6-624b718e5d00","Type":"ContainerDied","Data":"1223a99635ba79bdbf12681ff3637ec82020e6401d11d6095f9e9e76bae75ae9"} Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.072976 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1223a99635ba79bdbf12681ff3637ec82020e6401d11d6095f9e9e76bae75ae9" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.073057 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.102600 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/719c50b9-02ed-4138-8ce6-624b718e5d00-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.102630 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nslz5\" (UniqueName: \"kubernetes.io/projected/719c50b9-02ed-4138-8ce6-624b718e5d00-kube-api-access-nslz5\") on node \"crc\" DevicePath \"\"" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.102659 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/719c50b9-02ed-4138-8ce6-624b718e5d00-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.152807 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8"] Mar 21 08:58:35 crc kubenswrapper[4696]: E0321 08:58:35.153362 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="719c50b9-02ed-4138-8ce6-624b718e5d00" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.153387 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="719c50b9-02ed-4138-8ce6-624b718e5d00" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 21 08:58:35 crc kubenswrapper[4696]: E0321 08:58:35.153411 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6da51675-21b7-4468-b86f-28e19ec86020" containerName="oc" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.153420 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="6da51675-21b7-4468-b86f-28e19ec86020" containerName="oc" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.153658 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="719c50b9-02ed-4138-8ce6-624b718e5d00" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.153688 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="6da51675-21b7-4468-b86f-28e19ec86020" containerName="oc" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.154722 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.157375 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.157740 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bqp9r" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.157942 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.158057 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.165277 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8"] Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.306194 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxknk\" (UniqueName: \"kubernetes.io/projected/a8cdd5da-1ba0-4034-8a15-e3e86352772f-kube-api-access-kxknk\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8\" (UID: \"a8cdd5da-1ba0-4034-8a15-e3e86352772f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.306377 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8cdd5da-1ba0-4034-8a15-e3e86352772f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8\" (UID: \"a8cdd5da-1ba0-4034-8a15-e3e86352772f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.306458 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a8cdd5da-1ba0-4034-8a15-e3e86352772f-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8\" (UID: \"a8cdd5da-1ba0-4034-8a15-e3e86352772f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.408631 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxknk\" (UniqueName: \"kubernetes.io/projected/a8cdd5da-1ba0-4034-8a15-e3e86352772f-kube-api-access-kxknk\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8\" (UID: \"a8cdd5da-1ba0-4034-8a15-e3e86352772f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.408981 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8cdd5da-1ba0-4034-8a15-e3e86352772f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8\" (UID: \"a8cdd5da-1ba0-4034-8a15-e3e86352772f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.409127 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a8cdd5da-1ba0-4034-8a15-e3e86352772f-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8\" (UID: \"a8cdd5da-1ba0-4034-8a15-e3e86352772f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.412616 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a8cdd5da-1ba0-4034-8a15-e3e86352772f-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8\" (UID: \"a8cdd5da-1ba0-4034-8a15-e3e86352772f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.422683 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8cdd5da-1ba0-4034-8a15-e3e86352772f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8\" (UID: \"a8cdd5da-1ba0-4034-8a15-e3e86352772f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.428938 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxknk\" (UniqueName: \"kubernetes.io/projected/a8cdd5da-1ba0-4034-8a15-e3e86352772f-kube-api-access-kxknk\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8\" (UID: \"a8cdd5da-1ba0-4034-8a15-e3e86352772f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8" Mar 21 08:58:35 crc kubenswrapper[4696]: I0321 08:58:35.471289 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8" Mar 21 08:58:36 crc kubenswrapper[4696]: I0321 08:58:36.049065 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-297e-account-create-update-9rqlh"] Mar 21 08:58:36 crc kubenswrapper[4696]: I0321 08:58:36.060770 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-4t9zv"] Mar 21 08:58:36 crc kubenswrapper[4696]: I0321 08:58:36.085432 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-9b07-account-create-update-g6ddw"] Mar 21 08:58:36 crc kubenswrapper[4696]: I0321 08:58:36.089873 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8" event={"ID":"a8cdd5da-1ba0-4034-8a15-e3e86352772f","Type":"ContainerStarted","Data":"90b3a4e2a653d93c5603af9d1315c731a317ddf8b9539d717734ab51289847cb"} Mar 21 08:58:36 crc kubenswrapper[4696]: I0321 08:58:36.102908 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-8w2b4"] Mar 21 08:58:36 crc kubenswrapper[4696]: I0321 08:58:36.114118 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-9172-account-create-update-r4fhw"] Mar 21 08:58:36 crc kubenswrapper[4696]: I0321 08:58:36.124799 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-297e-account-create-update-9rqlh"] Mar 21 08:58:36 crc kubenswrapper[4696]: I0321 08:58:36.134338 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-p6m2c"] Mar 21 08:58:36 crc kubenswrapper[4696]: I0321 08:58:36.144752 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-9b07-account-create-update-g6ddw"] Mar 21 08:58:36 crc kubenswrapper[4696]: I0321 08:58:36.154150 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-4t9zv"] Mar 21 08:58:36 crc kubenswrapper[4696]: I0321 08:58:36.163890 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-p6m2c"] Mar 21 08:58:36 crc kubenswrapper[4696]: I0321 08:58:36.175876 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8"] Mar 21 08:58:36 crc kubenswrapper[4696]: I0321 08:58:36.221201 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-8w2b4"] Mar 21 08:58:36 crc kubenswrapper[4696]: I0321 08:58:36.232330 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-9172-account-create-update-r4fhw"] Mar 21 08:58:36 crc kubenswrapper[4696]: I0321 08:58:36.536093 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:58:36 crc kubenswrapper[4696]: E0321 08:58:36.536776 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:58:36 crc kubenswrapper[4696]: I0321 08:58:36.570749 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="092019b5-fac9-448e-be38-7b1cbc62f107" path="/var/lib/kubelet/pods/092019b5-fac9-448e-be38-7b1cbc62f107/volumes" Mar 21 08:58:36 crc kubenswrapper[4696]: I0321 08:58:36.578019 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d" path="/var/lib/kubelet/pods/2aefd2d5-17ac-43ef-b7ac-a06f7b46ed6d/volumes" Mar 21 08:58:36 crc kubenswrapper[4696]: I0321 08:58:36.578759 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54341261-11d8-407a-a339-6c0d70c1ad5f" path="/var/lib/kubelet/pods/54341261-11d8-407a-a339-6c0d70c1ad5f/volumes" Mar 21 08:58:36 crc kubenswrapper[4696]: I0321 08:58:36.579472 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="582c3a4a-774e-4748-b055-6e2dcf1065af" path="/var/lib/kubelet/pods/582c3a4a-774e-4748-b055-6e2dcf1065af/volumes" Mar 21 08:58:36 crc kubenswrapper[4696]: I0321 08:58:36.585931 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2" path="/var/lib/kubelet/pods/61c9ce3f-cf2b-4090-9bed-c7253d3ff4b2/volumes" Mar 21 08:58:36 crc kubenswrapper[4696]: I0321 08:58:36.586675 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9fe5e80-cc65-4225-9b74-73bf39a0f187" path="/var/lib/kubelet/pods/b9fe5e80-cc65-4225-9b74-73bf39a0f187/volumes" Mar 21 08:58:37 crc kubenswrapper[4696]: I0321 08:58:37.100619 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8" event={"ID":"a8cdd5da-1ba0-4034-8a15-e3e86352772f","Type":"ContainerStarted","Data":"115179da4c9aca59b027f7acf24d04656bd4dbcd8a532e6ae5ed297694358ad5"} Mar 21 08:58:37 crc kubenswrapper[4696]: I0321 08:58:37.121213 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8" podStartSLOduration=1.725193125 podStartE2EDuration="2.121195228s" podCreationTimestamp="2026-03-21 08:58:35 +0000 UTC" firstStartedPulling="2026-03-21 08:58:36.068225229 +0000 UTC m=+1850.189105942" lastFinishedPulling="2026-03-21 08:58:36.464227332 +0000 UTC m=+1850.585108045" observedRunningTime="2026-03-21 08:58:37.116957931 +0000 UTC m=+1851.237838654" watchObservedRunningTime="2026-03-21 08:58:37.121195228 +0000 UTC m=+1851.242075941" Mar 21 08:58:42 crc kubenswrapper[4696]: I0321 08:58:42.144917 4696 generic.go:334] "Generic (PLEG): container finished" podID="a8cdd5da-1ba0-4034-8a15-e3e86352772f" containerID="115179da4c9aca59b027f7acf24d04656bd4dbcd8a532e6ae5ed297694358ad5" exitCode=0 Mar 21 08:58:42 crc kubenswrapper[4696]: I0321 08:58:42.145003 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8" event={"ID":"a8cdd5da-1ba0-4034-8a15-e3e86352772f","Type":"ContainerDied","Data":"115179da4c9aca59b027f7acf24d04656bd4dbcd8a532e6ae5ed297694358ad5"} Mar 21 08:58:44 crc kubenswrapper[4696]: I0321 08:58:44.329996 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8" Mar 21 08:58:44 crc kubenswrapper[4696]: I0321 08:58:44.405857 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxknk\" (UniqueName: \"kubernetes.io/projected/a8cdd5da-1ba0-4034-8a15-e3e86352772f-kube-api-access-kxknk\") pod \"a8cdd5da-1ba0-4034-8a15-e3e86352772f\" (UID: \"a8cdd5da-1ba0-4034-8a15-e3e86352772f\") " Mar 21 08:58:44 crc kubenswrapper[4696]: I0321 08:58:44.406127 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a8cdd5da-1ba0-4034-8a15-e3e86352772f-ssh-key-openstack-edpm-ipam\") pod \"a8cdd5da-1ba0-4034-8a15-e3e86352772f\" (UID: \"a8cdd5da-1ba0-4034-8a15-e3e86352772f\") " Mar 21 08:58:44 crc kubenswrapper[4696]: I0321 08:58:44.406217 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8cdd5da-1ba0-4034-8a15-e3e86352772f-inventory\") pod \"a8cdd5da-1ba0-4034-8a15-e3e86352772f\" (UID: \"a8cdd5da-1ba0-4034-8a15-e3e86352772f\") " Mar 21 08:58:44 crc kubenswrapper[4696]: I0321 08:58:44.418078 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8cdd5da-1ba0-4034-8a15-e3e86352772f-kube-api-access-kxknk" (OuterVolumeSpecName: "kube-api-access-kxknk") pod "a8cdd5da-1ba0-4034-8a15-e3e86352772f" (UID: "a8cdd5da-1ba0-4034-8a15-e3e86352772f"). InnerVolumeSpecName "kube-api-access-kxknk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:58:44 crc kubenswrapper[4696]: I0321 08:58:44.450812 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8cdd5da-1ba0-4034-8a15-e3e86352772f-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a8cdd5da-1ba0-4034-8a15-e3e86352772f" (UID: "a8cdd5da-1ba0-4034-8a15-e3e86352772f"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:58:44 crc kubenswrapper[4696]: I0321 08:58:44.454234 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8cdd5da-1ba0-4034-8a15-e3e86352772f-inventory" (OuterVolumeSpecName: "inventory") pod "a8cdd5da-1ba0-4034-8a15-e3e86352772f" (UID: "a8cdd5da-1ba0-4034-8a15-e3e86352772f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:58:44 crc kubenswrapper[4696]: I0321 08:58:44.508913 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a8cdd5da-1ba0-4034-8a15-e3e86352772f-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 08:58:44 crc kubenswrapper[4696]: I0321 08:58:44.508951 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8cdd5da-1ba0-4034-8a15-e3e86352772f-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 08:58:44 crc kubenswrapper[4696]: I0321 08:58:44.508961 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxknk\" (UniqueName: \"kubernetes.io/projected/a8cdd5da-1ba0-4034-8a15-e3e86352772f-kube-api-access-kxknk\") on node \"crc\" DevicePath \"\"" Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.172692 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8" event={"ID":"a8cdd5da-1ba0-4034-8a15-e3e86352772f","Type":"ContainerDied","Data":"90b3a4e2a653d93c5603af9d1315c731a317ddf8b9539d717734ab51289847cb"} Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.173039 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90b3a4e2a653d93c5603af9d1315c731a317ddf8b9539d717734ab51289847cb" Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.172724 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8" Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.440798 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-85v2z"] Mar 21 08:58:45 crc kubenswrapper[4696]: E0321 08:58:45.441255 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8cdd5da-1ba0-4034-8a15-e3e86352772f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.441268 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8cdd5da-1ba0-4034-8a15-e3e86352772f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.441442 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8cdd5da-1ba0-4034-8a15-e3e86352772f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.442239 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-85v2z" Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.446004 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.446193 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.446218 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bqp9r" Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.446399 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.469084 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-85v2z"] Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.529429 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2v6tz\" (UniqueName: \"kubernetes.io/projected/3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd-kube-api-access-2v6tz\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-85v2z\" (UID: \"3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-85v2z" Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.529513 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-85v2z\" (UID: \"3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-85v2z" Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.529535 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-85v2z\" (UID: \"3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-85v2z" Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.631343 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2v6tz\" (UniqueName: \"kubernetes.io/projected/3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd-kube-api-access-2v6tz\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-85v2z\" (UID: \"3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-85v2z" Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.631454 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-85v2z\" (UID: \"3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-85v2z" Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.631484 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-85v2z\" (UID: \"3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-85v2z" Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.637418 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-85v2z\" (UID: \"3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-85v2z" Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.637924 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-85v2z\" (UID: \"3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-85v2z" Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.649436 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2v6tz\" (UniqueName: \"kubernetes.io/projected/3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd-kube-api-access-2v6tz\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-85v2z\" (UID: \"3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-85v2z" Mar 21 08:58:45 crc kubenswrapper[4696]: I0321 08:58:45.785428 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-85v2z" Mar 21 08:58:46 crc kubenswrapper[4696]: I0321 08:58:46.569663 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-85v2z"] Mar 21 08:58:47 crc kubenswrapper[4696]: I0321 08:58:47.010978 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 08:58:47 crc kubenswrapper[4696]: I0321 08:58:47.190591 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-85v2z" event={"ID":"3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd","Type":"ContainerStarted","Data":"fd504b1a581c5d28c9a3ad88e1154ca7cc2d4400e524bcbd695fca21c3ea54d5"} Mar 21 08:58:48 crc kubenswrapper[4696]: I0321 08:58:48.202466 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-85v2z" event={"ID":"3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd","Type":"ContainerStarted","Data":"3ce44330d7fd4f843fed115d83807c3e6f6ddf25a9281ef54c71983ce94e7813"} Mar 21 08:58:48 crc kubenswrapper[4696]: I0321 08:58:48.227109 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-85v2z" podStartSLOduration=2.787188318 podStartE2EDuration="3.227089104s" podCreationTimestamp="2026-03-21 08:58:45 +0000 UTC" firstStartedPulling="2026-03-21 08:58:46.568796216 +0000 UTC m=+1860.689676929" lastFinishedPulling="2026-03-21 08:58:47.008697002 +0000 UTC m=+1861.129577715" observedRunningTime="2026-03-21 08:58:48.218087487 +0000 UTC m=+1862.338968200" watchObservedRunningTime="2026-03-21 08:58:48.227089104 +0000 UTC m=+1862.347969817" Mar 21 08:58:51 crc kubenswrapper[4696]: I0321 08:58:51.534593 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:58:51 crc kubenswrapper[4696]: E0321 08:58:51.535259 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:59:00 crc kubenswrapper[4696]: I0321 08:59:00.372216 4696 scope.go:117] "RemoveContainer" containerID="115b17be47f80234d621a6e9d263a7d4f73af5b708364bcc07917182ddced170" Mar 21 08:59:00 crc kubenswrapper[4696]: I0321 08:59:00.403751 4696 scope.go:117] "RemoveContainer" containerID="a01d9d82b3cfea30bfced4f72a1e28c6089568366cfb2d31ad682c7a39cec486" Mar 21 08:59:00 crc kubenswrapper[4696]: I0321 08:59:00.453290 4696 scope.go:117] "RemoveContainer" containerID="2e2b074e07223bf1072106b86484d2419aec5ad3e84801171bcc9236fe9e7473" Mar 21 08:59:00 crc kubenswrapper[4696]: I0321 08:59:00.495966 4696 scope.go:117] "RemoveContainer" containerID="f11f6249b47a767958fabaa0043b96567dd8ae1f79f4f4bb03f10526f851234f" Mar 21 08:59:00 crc kubenswrapper[4696]: I0321 08:59:00.551557 4696 scope.go:117] "RemoveContainer" containerID="d691cb047a6743e59099baef4ff05d5f1de76cddf6715b786b4e07e0cc1eb607" Mar 21 08:59:00 crc kubenswrapper[4696]: I0321 08:59:00.605216 4696 scope.go:117] "RemoveContainer" containerID="c90e0fb672ad228ae8aaa4a4b85436ed85fdd11eefb9c515d4a45dca0ae81b53" Mar 21 08:59:00 crc kubenswrapper[4696]: I0321 08:59:00.665266 4696 scope.go:117] "RemoveContainer" containerID="499e153025457892a2d3dd1ae7d23b19b1557d52982f153cdf92e437caa1cd13" Mar 21 08:59:00 crc kubenswrapper[4696]: I0321 08:59:00.707405 4696 scope.go:117] "RemoveContainer" containerID="b3d0f17f9ae93f081735e0b7f189d1fc592d83b4137d6f0771c485ed6c35106d" Mar 21 08:59:04 crc kubenswrapper[4696]: I0321 08:59:04.540906 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:59:04 crc kubenswrapper[4696]: E0321 08:59:04.541737 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:59:13 crc kubenswrapper[4696]: I0321 08:59:13.050057 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-j298g"] Mar 21 08:59:13 crc kubenswrapper[4696]: I0321 08:59:13.062577 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-j298g"] Mar 21 08:59:14 crc kubenswrapper[4696]: I0321 08:59:14.545471 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e448486-5596-4442-a11a-804a7b7b94d8" path="/var/lib/kubelet/pods/5e448486-5596-4442-a11a-804a7b7b94d8/volumes" Mar 21 08:59:17 crc kubenswrapper[4696]: I0321 08:59:17.534216 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:59:17 crc kubenswrapper[4696]: E0321 08:59:17.535171 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:59:21 crc kubenswrapper[4696]: I0321 08:59:21.520964 4696 generic.go:334] "Generic (PLEG): container finished" podID="3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd" containerID="3ce44330d7fd4f843fed115d83807c3e6f6ddf25a9281ef54c71983ce94e7813" exitCode=0 Mar 21 08:59:21 crc kubenswrapper[4696]: I0321 08:59:21.521052 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-85v2z" event={"ID":"3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd","Type":"ContainerDied","Data":"3ce44330d7fd4f843fed115d83807c3e6f6ddf25a9281ef54c71983ce94e7813"} Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.491974 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-85v2z" Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.554459 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-85v2z" event={"ID":"3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd","Type":"ContainerDied","Data":"fd504b1a581c5d28c9a3ad88e1154ca7cc2d4400e524bcbd695fca21c3ea54d5"} Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.554512 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd504b1a581c5d28c9a3ad88e1154ca7cc2d4400e524bcbd695fca21c3ea54d5" Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.554581 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-85v2z" Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.626115 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6k46h"] Mar 21 08:59:23 crc kubenswrapper[4696]: E0321 08:59:23.626854 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.626872 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.627046 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.627853 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6k46h" Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.631874 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd-ssh-key-openstack-edpm-ipam\") pod \"3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd\" (UID: \"3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd\") " Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.631950 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd-inventory\") pod \"3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd\" (UID: \"3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd\") " Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.631979 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2v6tz\" (UniqueName: \"kubernetes.io/projected/3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd-kube-api-access-2v6tz\") pod \"3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd\" (UID: \"3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd\") " Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.638724 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd-kube-api-access-2v6tz" (OuterVolumeSpecName: "kube-api-access-2v6tz") pod "3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd" (UID: "3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd"). InnerVolumeSpecName "kube-api-access-2v6tz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.644103 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6k46h"] Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.671222 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd" (UID: "3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.677246 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd-inventory" (OuterVolumeSpecName: "inventory") pod "3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd" (UID: "3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.735888 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qw47\" (UniqueName: \"kubernetes.io/projected/7f2cc308-7143-4847-9e6d-a96e9ddd6c51-kube-api-access-5qw47\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6k46h\" (UID: \"7f2cc308-7143-4847-9e6d-a96e9ddd6c51\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6k46h" Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.735984 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f2cc308-7143-4847-9e6d-a96e9ddd6c51-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6k46h\" (UID: \"7f2cc308-7143-4847-9e6d-a96e9ddd6c51\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6k46h" Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.736867 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7f2cc308-7143-4847-9e6d-a96e9ddd6c51-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6k46h\" (UID: \"7f2cc308-7143-4847-9e6d-a96e9ddd6c51\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6k46h" Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.737096 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.737118 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.737130 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2v6tz\" (UniqueName: \"kubernetes.io/projected/3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd-kube-api-access-2v6tz\") on node \"crc\" DevicePath \"\"" Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.839165 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qw47\" (UniqueName: \"kubernetes.io/projected/7f2cc308-7143-4847-9e6d-a96e9ddd6c51-kube-api-access-5qw47\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6k46h\" (UID: \"7f2cc308-7143-4847-9e6d-a96e9ddd6c51\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6k46h" Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.839538 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f2cc308-7143-4847-9e6d-a96e9ddd6c51-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6k46h\" (UID: \"7f2cc308-7143-4847-9e6d-a96e9ddd6c51\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6k46h" Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.839768 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7f2cc308-7143-4847-9e6d-a96e9ddd6c51-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6k46h\" (UID: \"7f2cc308-7143-4847-9e6d-a96e9ddd6c51\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6k46h" Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.843221 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f2cc308-7143-4847-9e6d-a96e9ddd6c51-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6k46h\" (UID: \"7f2cc308-7143-4847-9e6d-a96e9ddd6c51\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6k46h" Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.844045 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7f2cc308-7143-4847-9e6d-a96e9ddd6c51-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6k46h\" (UID: \"7f2cc308-7143-4847-9e6d-a96e9ddd6c51\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6k46h" Mar 21 08:59:23 crc kubenswrapper[4696]: I0321 08:59:23.857679 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qw47\" (UniqueName: \"kubernetes.io/projected/7f2cc308-7143-4847-9e6d-a96e9ddd6c51-kube-api-access-5qw47\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6k46h\" (UID: \"7f2cc308-7143-4847-9e6d-a96e9ddd6c51\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6k46h" Mar 21 08:59:24 crc kubenswrapper[4696]: I0321 08:59:24.050919 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6k46h" Mar 21 08:59:24 crc kubenswrapper[4696]: I0321 08:59:24.546967 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6k46h"] Mar 21 08:59:24 crc kubenswrapper[4696]: I0321 08:59:24.564314 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6k46h" event={"ID":"7f2cc308-7143-4847-9e6d-a96e9ddd6c51","Type":"ContainerStarted","Data":"7f7ba14296f05775fce3272f66a4bd609bf7c5294d38b04969f5d3ece62b59e1"} Mar 21 08:59:25 crc kubenswrapper[4696]: I0321 08:59:25.576537 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6k46h" event={"ID":"7f2cc308-7143-4847-9e6d-a96e9ddd6c51","Type":"ContainerStarted","Data":"8dbe970a78a2cc2846ab38dd58e4bd1842de50bd19507c7a3a91aec844ef64e7"} Mar 21 08:59:25 crc kubenswrapper[4696]: I0321 08:59:25.609287 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6k46h" podStartSLOduration=2.037242454 podStartE2EDuration="2.609261433s" podCreationTimestamp="2026-03-21 08:59:23 +0000 UTC" firstStartedPulling="2026-03-21 08:59:24.534493183 +0000 UTC m=+1898.655373896" lastFinishedPulling="2026-03-21 08:59:25.106512162 +0000 UTC m=+1899.227392875" observedRunningTime="2026-03-21 08:59:25.594559971 +0000 UTC m=+1899.715440704" watchObservedRunningTime="2026-03-21 08:59:25.609261433 +0000 UTC m=+1899.730142176" Mar 21 08:59:29 crc kubenswrapper[4696]: I0321 08:59:29.547268 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:59:29 crc kubenswrapper[4696]: E0321 08:59:29.548077 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:59:36 crc kubenswrapper[4696]: I0321 08:59:36.041625 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-d4v88"] Mar 21 08:59:36 crc kubenswrapper[4696]: I0321 08:59:36.053630 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-d4v88"] Mar 21 08:59:36 crc kubenswrapper[4696]: I0321 08:59:36.547458 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29e6fe18-05b8-430f-b3fe-96ca1cf04a94" path="/var/lib/kubelet/pods/29e6fe18-05b8-430f-b3fe-96ca1cf04a94/volumes" Mar 21 08:59:40 crc kubenswrapper[4696]: I0321 08:59:40.030736 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fjcrs"] Mar 21 08:59:40 crc kubenswrapper[4696]: I0321 08:59:40.039611 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fjcrs"] Mar 21 08:59:40 crc kubenswrapper[4696]: I0321 08:59:40.547272 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4856e60e-c40e-4b95-bc64-2f33b8b79e4b" path="/var/lib/kubelet/pods/4856e60e-c40e-4b95-bc64-2f33b8b79e4b/volumes" Mar 21 08:59:41 crc kubenswrapper[4696]: I0321 08:59:41.535192 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:59:41 crc kubenswrapper[4696]: E0321 08:59:41.535770 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 08:59:56 crc kubenswrapper[4696]: I0321 08:59:56.545384 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 08:59:56 crc kubenswrapper[4696]: E0321 08:59:56.546619 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.148092 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568060-qbn8d"] Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.150058 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568060-qbn8d" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.151891 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.152267 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.152777 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.160741 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568060-bj9q2"] Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.162789 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568060-bj9q2" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.174316 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.174370 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.195696 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568060-qbn8d"] Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.215989 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568060-bj9q2"] Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.340926 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cggft\" (UniqueName: \"kubernetes.io/projected/abbe55ca-24b3-46de-a978-aa00abe0d9e9-kube-api-access-cggft\") pod \"collect-profiles-29568060-bj9q2\" (UID: \"abbe55ca-24b3-46de-a978-aa00abe0d9e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568060-bj9q2" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.341344 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfx6t\" (UniqueName: \"kubernetes.io/projected/140b0556-76e9-4138-9395-3c7bacf16cb2-kube-api-access-lfx6t\") pod \"auto-csr-approver-29568060-qbn8d\" (UID: \"140b0556-76e9-4138-9395-3c7bacf16cb2\") " pod="openshift-infra/auto-csr-approver-29568060-qbn8d" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.341514 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/abbe55ca-24b3-46de-a978-aa00abe0d9e9-config-volume\") pod \"collect-profiles-29568060-bj9q2\" (UID: \"abbe55ca-24b3-46de-a978-aa00abe0d9e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568060-bj9q2" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.341692 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/abbe55ca-24b3-46de-a978-aa00abe0d9e9-secret-volume\") pod \"collect-profiles-29568060-bj9q2\" (UID: \"abbe55ca-24b3-46de-a978-aa00abe0d9e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568060-bj9q2" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.443540 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfx6t\" (UniqueName: \"kubernetes.io/projected/140b0556-76e9-4138-9395-3c7bacf16cb2-kube-api-access-lfx6t\") pod \"auto-csr-approver-29568060-qbn8d\" (UID: \"140b0556-76e9-4138-9395-3c7bacf16cb2\") " pod="openshift-infra/auto-csr-approver-29568060-qbn8d" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.443660 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/abbe55ca-24b3-46de-a978-aa00abe0d9e9-config-volume\") pod \"collect-profiles-29568060-bj9q2\" (UID: \"abbe55ca-24b3-46de-a978-aa00abe0d9e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568060-bj9q2" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.443791 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/abbe55ca-24b3-46de-a978-aa00abe0d9e9-secret-volume\") pod \"collect-profiles-29568060-bj9q2\" (UID: \"abbe55ca-24b3-46de-a978-aa00abe0d9e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568060-bj9q2" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.444114 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cggft\" (UniqueName: \"kubernetes.io/projected/abbe55ca-24b3-46de-a978-aa00abe0d9e9-kube-api-access-cggft\") pod \"collect-profiles-29568060-bj9q2\" (UID: \"abbe55ca-24b3-46de-a978-aa00abe0d9e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568060-bj9q2" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.444861 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/abbe55ca-24b3-46de-a978-aa00abe0d9e9-config-volume\") pod \"collect-profiles-29568060-bj9q2\" (UID: \"abbe55ca-24b3-46de-a978-aa00abe0d9e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568060-bj9q2" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.460365 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/abbe55ca-24b3-46de-a978-aa00abe0d9e9-secret-volume\") pod \"collect-profiles-29568060-bj9q2\" (UID: \"abbe55ca-24b3-46de-a978-aa00abe0d9e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568060-bj9q2" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.462679 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfx6t\" (UniqueName: \"kubernetes.io/projected/140b0556-76e9-4138-9395-3c7bacf16cb2-kube-api-access-lfx6t\") pod \"auto-csr-approver-29568060-qbn8d\" (UID: \"140b0556-76e9-4138-9395-3c7bacf16cb2\") " pod="openshift-infra/auto-csr-approver-29568060-qbn8d" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.471713 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cggft\" (UniqueName: \"kubernetes.io/projected/abbe55ca-24b3-46de-a978-aa00abe0d9e9-kube-api-access-cggft\") pod \"collect-profiles-29568060-bj9q2\" (UID: \"abbe55ca-24b3-46de-a978-aa00abe0d9e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568060-bj9q2" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.476681 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568060-qbn8d" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.501907 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568060-bj9q2" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.896558 4696 scope.go:117] "RemoveContainer" containerID="9eae8c5cbd9fc1b2ae5e65811cb3ddc5f1961c88f5db9e6c115709c50c3f3c03" Mar 21 09:00:00 crc kubenswrapper[4696]: I0321 09:00:00.960926 4696 scope.go:117] "RemoveContainer" containerID="4be79300257c4c766e3f6f08ee86a69724e964c534ae69af90e17eeac61e8aa6" Mar 21 09:00:01 crc kubenswrapper[4696]: I0321 09:00:01.002601 4696 scope.go:117] "RemoveContainer" containerID="32c9d7bf028548a11c13896a74b2d7e729721772af461048e7575f7de03057ed" Mar 21 09:00:01 crc kubenswrapper[4696]: I0321 09:00:01.182082 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568060-qbn8d"] Mar 21 09:00:01 crc kubenswrapper[4696]: W0321 09:00:01.188024 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podabbe55ca_24b3_46de_a978_aa00abe0d9e9.slice/crio-3536686d77493934345d9520c7eae69c3ccacfafa1029ccdded367098c1f4ec1 WatchSource:0}: Error finding container 3536686d77493934345d9520c7eae69c3ccacfafa1029ccdded367098c1f4ec1: Status 404 returned error can't find the container with id 3536686d77493934345d9520c7eae69c3ccacfafa1029ccdded367098c1f4ec1 Mar 21 09:00:01 crc kubenswrapper[4696]: W0321 09:00:01.188391 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod140b0556_76e9_4138_9395_3c7bacf16cb2.slice/crio-d6d66ad1096bc2fb7bbcdcde03ebc3d6cc5dc61a74175152fa91f31d87479486 WatchSource:0}: Error finding container d6d66ad1096bc2fb7bbcdcde03ebc3d6cc5dc61a74175152fa91f31d87479486: Status 404 returned error can't find the container with id d6d66ad1096bc2fb7bbcdcde03ebc3d6cc5dc61a74175152fa91f31d87479486 Mar 21 09:00:01 crc kubenswrapper[4696]: I0321 09:00:01.195210 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568060-bj9q2"] Mar 21 09:00:01 crc kubenswrapper[4696]: I0321 09:00:01.922403 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568060-qbn8d" event={"ID":"140b0556-76e9-4138-9395-3c7bacf16cb2","Type":"ContainerStarted","Data":"d6d66ad1096bc2fb7bbcdcde03ebc3d6cc5dc61a74175152fa91f31d87479486"} Mar 21 09:00:01 crc kubenswrapper[4696]: I0321 09:00:01.924743 4696 generic.go:334] "Generic (PLEG): container finished" podID="abbe55ca-24b3-46de-a978-aa00abe0d9e9" containerID="851bdfdd749137408d17275f09bb435667b969c024449803c98cfaf952273c1d" exitCode=0 Mar 21 09:00:01 crc kubenswrapper[4696]: I0321 09:00:01.924771 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568060-bj9q2" event={"ID":"abbe55ca-24b3-46de-a978-aa00abe0d9e9","Type":"ContainerDied","Data":"851bdfdd749137408d17275f09bb435667b969c024449803c98cfaf952273c1d"} Mar 21 09:00:01 crc kubenswrapper[4696]: I0321 09:00:01.924848 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568060-bj9q2" event={"ID":"abbe55ca-24b3-46de-a978-aa00abe0d9e9","Type":"ContainerStarted","Data":"3536686d77493934345d9520c7eae69c3ccacfafa1029ccdded367098c1f4ec1"} Mar 21 09:00:03 crc kubenswrapper[4696]: I0321 09:00:03.739988 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568060-bj9q2" Mar 21 09:00:03 crc kubenswrapper[4696]: I0321 09:00:03.819937 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/abbe55ca-24b3-46de-a978-aa00abe0d9e9-secret-volume\") pod \"abbe55ca-24b3-46de-a978-aa00abe0d9e9\" (UID: \"abbe55ca-24b3-46de-a978-aa00abe0d9e9\") " Mar 21 09:00:03 crc kubenswrapper[4696]: I0321 09:00:03.821191 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cggft\" (UniqueName: \"kubernetes.io/projected/abbe55ca-24b3-46de-a978-aa00abe0d9e9-kube-api-access-cggft\") pod \"abbe55ca-24b3-46de-a978-aa00abe0d9e9\" (UID: \"abbe55ca-24b3-46de-a978-aa00abe0d9e9\") " Mar 21 09:00:03 crc kubenswrapper[4696]: I0321 09:00:03.821267 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/abbe55ca-24b3-46de-a978-aa00abe0d9e9-config-volume\") pod \"abbe55ca-24b3-46de-a978-aa00abe0d9e9\" (UID: \"abbe55ca-24b3-46de-a978-aa00abe0d9e9\") " Mar 21 09:00:03 crc kubenswrapper[4696]: I0321 09:00:03.822128 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abbe55ca-24b3-46de-a978-aa00abe0d9e9-config-volume" (OuterVolumeSpecName: "config-volume") pod "abbe55ca-24b3-46de-a978-aa00abe0d9e9" (UID: "abbe55ca-24b3-46de-a978-aa00abe0d9e9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 09:00:03 crc kubenswrapper[4696]: I0321 09:00:03.826386 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abbe55ca-24b3-46de-a978-aa00abe0d9e9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "abbe55ca-24b3-46de-a978-aa00abe0d9e9" (UID: "abbe55ca-24b3-46de-a978-aa00abe0d9e9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:00:03 crc kubenswrapper[4696]: I0321 09:00:03.832581 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abbe55ca-24b3-46de-a978-aa00abe0d9e9-kube-api-access-cggft" (OuterVolumeSpecName: "kube-api-access-cggft") pod "abbe55ca-24b3-46de-a978-aa00abe0d9e9" (UID: "abbe55ca-24b3-46de-a978-aa00abe0d9e9"). InnerVolumeSpecName "kube-api-access-cggft". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:00:03 crc kubenswrapper[4696]: I0321 09:00:03.924030 4696 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/abbe55ca-24b3-46de-a978-aa00abe0d9e9-config-volume\") on node \"crc\" DevicePath \"\"" Mar 21 09:00:03 crc kubenswrapper[4696]: I0321 09:00:03.924293 4696 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/abbe55ca-24b3-46de-a978-aa00abe0d9e9-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 21 09:00:03 crc kubenswrapper[4696]: I0321 09:00:03.924357 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cggft\" (UniqueName: \"kubernetes.io/projected/abbe55ca-24b3-46de-a978-aa00abe0d9e9-kube-api-access-cggft\") on node \"crc\" DevicePath \"\"" Mar 21 09:00:03 crc kubenswrapper[4696]: I0321 09:00:03.947801 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568060-bj9q2" event={"ID":"abbe55ca-24b3-46de-a978-aa00abe0d9e9","Type":"ContainerDied","Data":"3536686d77493934345d9520c7eae69c3ccacfafa1029ccdded367098c1f4ec1"} Mar 21 09:00:03 crc kubenswrapper[4696]: I0321 09:00:03.948098 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3536686d77493934345d9520c7eae69c3ccacfafa1029ccdded367098c1f4ec1" Mar 21 09:00:03 crc kubenswrapper[4696]: I0321 09:00:03.947890 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568060-bj9q2" Mar 21 09:00:04 crc kubenswrapper[4696]: I0321 09:00:04.821727 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48"] Mar 21 09:00:04 crc kubenswrapper[4696]: I0321 09:00:04.831669 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568015-msx48"] Mar 21 09:00:04 crc kubenswrapper[4696]: I0321 09:00:04.961417 4696 generic.go:334] "Generic (PLEG): container finished" podID="140b0556-76e9-4138-9395-3c7bacf16cb2" containerID="8409d25fad0117a45a1a1c8ba21857376c1508984d7527330e44a7c31208f36e" exitCode=0 Mar 21 09:00:04 crc kubenswrapper[4696]: I0321 09:00:04.961475 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568060-qbn8d" event={"ID":"140b0556-76e9-4138-9395-3c7bacf16cb2","Type":"ContainerDied","Data":"8409d25fad0117a45a1a1c8ba21857376c1508984d7527330e44a7c31208f36e"} Mar 21 09:00:06 crc kubenswrapper[4696]: I0321 09:00:06.547165 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ef60c27-2062-4bc8-8c37-e13db97d29a3" path="/var/lib/kubelet/pods/6ef60c27-2062-4bc8-8c37-e13db97d29a3/volumes" Mar 21 09:00:06 crc kubenswrapper[4696]: I0321 09:00:06.638533 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568060-qbn8d" Mar 21 09:00:06 crc kubenswrapper[4696]: I0321 09:00:06.781997 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfx6t\" (UniqueName: \"kubernetes.io/projected/140b0556-76e9-4138-9395-3c7bacf16cb2-kube-api-access-lfx6t\") pod \"140b0556-76e9-4138-9395-3c7bacf16cb2\" (UID: \"140b0556-76e9-4138-9395-3c7bacf16cb2\") " Mar 21 09:00:06 crc kubenswrapper[4696]: I0321 09:00:06.788857 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/140b0556-76e9-4138-9395-3c7bacf16cb2-kube-api-access-lfx6t" (OuterVolumeSpecName: "kube-api-access-lfx6t") pod "140b0556-76e9-4138-9395-3c7bacf16cb2" (UID: "140b0556-76e9-4138-9395-3c7bacf16cb2"). InnerVolumeSpecName "kube-api-access-lfx6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:00:06 crc kubenswrapper[4696]: I0321 09:00:06.884440 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfx6t\" (UniqueName: \"kubernetes.io/projected/140b0556-76e9-4138-9395-3c7bacf16cb2-kube-api-access-lfx6t\") on node \"crc\" DevicePath \"\"" Mar 21 09:00:06 crc kubenswrapper[4696]: I0321 09:00:06.978696 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568060-qbn8d" event={"ID":"140b0556-76e9-4138-9395-3c7bacf16cb2","Type":"ContainerDied","Data":"d6d66ad1096bc2fb7bbcdcde03ebc3d6cc5dc61a74175152fa91f31d87479486"} Mar 21 09:00:06 crc kubenswrapper[4696]: I0321 09:00:06.978736 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6d66ad1096bc2fb7bbcdcde03ebc3d6cc5dc61a74175152fa91f31d87479486" Mar 21 09:00:06 crc kubenswrapper[4696]: I0321 09:00:06.978759 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568060-qbn8d" Mar 21 09:00:07 crc kubenswrapper[4696]: I0321 09:00:07.691459 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568054-l85jh"] Mar 21 09:00:07 crc kubenswrapper[4696]: I0321 09:00:07.702573 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568054-l85jh"] Mar 21 09:00:08 crc kubenswrapper[4696]: I0321 09:00:08.535133 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 09:00:08 crc kubenswrapper[4696]: I0321 09:00:08.546267 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03261136-a65c-455b-9e7d-f24d95e68322" path="/var/lib/kubelet/pods/03261136-a65c-455b-9e7d-f24d95e68322/volumes" Mar 21 09:00:08 crc kubenswrapper[4696]: I0321 09:00:08.997830 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerStarted","Data":"3187c3e490f50d1673fbeb2f54747e27d740a56d231e0d24142915c7db238efa"} Mar 21 09:00:09 crc kubenswrapper[4696]: I0321 09:00:09.001477 4696 generic.go:334] "Generic (PLEG): container finished" podID="7f2cc308-7143-4847-9e6d-a96e9ddd6c51" containerID="8dbe970a78a2cc2846ab38dd58e4bd1842de50bd19507c7a3a91aec844ef64e7" exitCode=0 Mar 21 09:00:09 crc kubenswrapper[4696]: I0321 09:00:09.001518 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6k46h" event={"ID":"7f2cc308-7143-4847-9e6d-a96e9ddd6c51","Type":"ContainerDied","Data":"8dbe970a78a2cc2846ab38dd58e4bd1842de50bd19507c7a3a91aec844ef64e7"} Mar 21 09:00:10 crc kubenswrapper[4696]: I0321 09:00:10.769402 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6k46h" Mar 21 09:00:10 crc kubenswrapper[4696]: I0321 09:00:10.866087 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7f2cc308-7143-4847-9e6d-a96e9ddd6c51-ssh-key-openstack-edpm-ipam\") pod \"7f2cc308-7143-4847-9e6d-a96e9ddd6c51\" (UID: \"7f2cc308-7143-4847-9e6d-a96e9ddd6c51\") " Mar 21 09:00:10 crc kubenswrapper[4696]: I0321 09:00:10.866219 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qw47\" (UniqueName: \"kubernetes.io/projected/7f2cc308-7143-4847-9e6d-a96e9ddd6c51-kube-api-access-5qw47\") pod \"7f2cc308-7143-4847-9e6d-a96e9ddd6c51\" (UID: \"7f2cc308-7143-4847-9e6d-a96e9ddd6c51\") " Mar 21 09:00:10 crc kubenswrapper[4696]: I0321 09:00:10.866368 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f2cc308-7143-4847-9e6d-a96e9ddd6c51-inventory\") pod \"7f2cc308-7143-4847-9e6d-a96e9ddd6c51\" (UID: \"7f2cc308-7143-4847-9e6d-a96e9ddd6c51\") " Mar 21 09:00:10 crc kubenswrapper[4696]: I0321 09:00:10.872314 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f2cc308-7143-4847-9e6d-a96e9ddd6c51-kube-api-access-5qw47" (OuterVolumeSpecName: "kube-api-access-5qw47") pod "7f2cc308-7143-4847-9e6d-a96e9ddd6c51" (UID: "7f2cc308-7143-4847-9e6d-a96e9ddd6c51"). InnerVolumeSpecName "kube-api-access-5qw47". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:00:10 crc kubenswrapper[4696]: I0321 09:00:10.911363 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f2cc308-7143-4847-9e6d-a96e9ddd6c51-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "7f2cc308-7143-4847-9e6d-a96e9ddd6c51" (UID: "7f2cc308-7143-4847-9e6d-a96e9ddd6c51"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:00:10 crc kubenswrapper[4696]: I0321 09:00:10.916096 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f2cc308-7143-4847-9e6d-a96e9ddd6c51-inventory" (OuterVolumeSpecName: "inventory") pod "7f2cc308-7143-4847-9e6d-a96e9ddd6c51" (UID: "7f2cc308-7143-4847-9e6d-a96e9ddd6c51"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:00:10 crc kubenswrapper[4696]: I0321 09:00:10.968591 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7f2cc308-7143-4847-9e6d-a96e9ddd6c51-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 09:00:10 crc kubenswrapper[4696]: I0321 09:00:10.968618 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qw47\" (UniqueName: \"kubernetes.io/projected/7f2cc308-7143-4847-9e6d-a96e9ddd6c51-kube-api-access-5qw47\") on node \"crc\" DevicePath \"\"" Mar 21 09:00:10 crc kubenswrapper[4696]: I0321 09:00:10.968631 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f2cc308-7143-4847-9e6d-a96e9ddd6c51-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.020360 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6k46h" event={"ID":"7f2cc308-7143-4847-9e6d-a96e9ddd6c51","Type":"ContainerDied","Data":"7f7ba14296f05775fce3272f66a4bd609bf7c5294d38b04969f5d3ece62b59e1"} Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.020403 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f7ba14296f05775fce3272f66a4bd609bf7c5294d38b04969f5d3ece62b59e1" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.020382 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6k46h" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.132942 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ckxp8"] Mar 21 09:00:11 crc kubenswrapper[4696]: E0321 09:00:11.133637 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abbe55ca-24b3-46de-a978-aa00abe0d9e9" containerName="collect-profiles" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.133650 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="abbe55ca-24b3-46de-a978-aa00abe0d9e9" containerName="collect-profiles" Mar 21 09:00:11 crc kubenswrapper[4696]: E0321 09:00:11.133665 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f2cc308-7143-4847-9e6d-a96e9ddd6c51" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.133673 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f2cc308-7143-4847-9e6d-a96e9ddd6c51" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 21 09:00:11 crc kubenswrapper[4696]: E0321 09:00:11.133687 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="140b0556-76e9-4138-9395-3c7bacf16cb2" containerName="oc" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.133693 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="140b0556-76e9-4138-9395-3c7bacf16cb2" containerName="oc" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.133959 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="abbe55ca-24b3-46de-a978-aa00abe0d9e9" containerName="collect-profiles" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.133983 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="140b0556-76e9-4138-9395-3c7bacf16cb2" containerName="oc" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.133993 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f2cc308-7143-4847-9e6d-a96e9ddd6c51" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.134913 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ckxp8" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.139487 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bqp9r" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.139751 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.139934 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.140112 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.143870 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ckxp8"] Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.276080 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ckxp8\" (UID: \"4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6\") " pod="openstack/ssh-known-hosts-edpm-deployment-ckxp8" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.276183 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqmpj\" (UniqueName: \"kubernetes.io/projected/4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6-kube-api-access-tqmpj\") pod \"ssh-known-hosts-edpm-deployment-ckxp8\" (UID: \"4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6\") " pod="openstack/ssh-known-hosts-edpm-deployment-ckxp8" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.276283 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ckxp8\" (UID: \"4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6\") " pod="openstack/ssh-known-hosts-edpm-deployment-ckxp8" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.378796 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ckxp8\" (UID: \"4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6\") " pod="openstack/ssh-known-hosts-edpm-deployment-ckxp8" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.378964 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqmpj\" (UniqueName: \"kubernetes.io/projected/4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6-kube-api-access-tqmpj\") pod \"ssh-known-hosts-edpm-deployment-ckxp8\" (UID: \"4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6\") " pod="openstack/ssh-known-hosts-edpm-deployment-ckxp8" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.379065 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ckxp8\" (UID: \"4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6\") " pod="openstack/ssh-known-hosts-edpm-deployment-ckxp8" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.384675 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ckxp8\" (UID: \"4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6\") " pod="openstack/ssh-known-hosts-edpm-deployment-ckxp8" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.385452 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ckxp8\" (UID: \"4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6\") " pod="openstack/ssh-known-hosts-edpm-deployment-ckxp8" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.395595 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqmpj\" (UniqueName: \"kubernetes.io/projected/4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6-kube-api-access-tqmpj\") pod \"ssh-known-hosts-edpm-deployment-ckxp8\" (UID: \"4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6\") " pod="openstack/ssh-known-hosts-edpm-deployment-ckxp8" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.466941 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ckxp8" Mar 21 09:00:11 crc kubenswrapper[4696]: I0321 09:00:11.996841 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ckxp8"] Mar 21 09:00:12 crc kubenswrapper[4696]: W0321 09:00:12.001145 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f9d8f9c_fde1_4198_b24e_bcffe8ace9b6.slice/crio-c6cb43a9b89a21714e3267d7ead5f94af02a4768f516cc543e83bedbd98ae7aa WatchSource:0}: Error finding container c6cb43a9b89a21714e3267d7ead5f94af02a4768f516cc543e83bedbd98ae7aa: Status 404 returned error can't find the container with id c6cb43a9b89a21714e3267d7ead5f94af02a4768f516cc543e83bedbd98ae7aa Mar 21 09:00:12 crc kubenswrapper[4696]: I0321 09:00:12.033459 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ckxp8" event={"ID":"4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6","Type":"ContainerStarted","Data":"c6cb43a9b89a21714e3267d7ead5f94af02a4768f516cc543e83bedbd98ae7aa"} Mar 21 09:00:13 crc kubenswrapper[4696]: I0321 09:00:13.047159 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ckxp8" event={"ID":"4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6","Type":"ContainerStarted","Data":"d3d2352f4c33b732be2fe386b7af0101a6d6789f554df03de05a694fe575b9e5"} Mar 21 09:00:13 crc kubenswrapper[4696]: I0321 09:00:13.074157 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-ckxp8" podStartSLOduration=1.6149919499999998 podStartE2EDuration="2.074136446s" podCreationTimestamp="2026-03-21 09:00:11 +0000 UTC" firstStartedPulling="2026-03-21 09:00:12.003553651 +0000 UTC m=+1946.124434384" lastFinishedPulling="2026-03-21 09:00:12.462698167 +0000 UTC m=+1946.583578880" observedRunningTime="2026-03-21 09:00:13.060500042 +0000 UTC m=+1947.181380755" watchObservedRunningTime="2026-03-21 09:00:13.074136446 +0000 UTC m=+1947.195017179" Mar 21 09:00:19 crc kubenswrapper[4696]: I0321 09:00:19.121084 4696 generic.go:334] "Generic (PLEG): container finished" podID="4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6" containerID="d3d2352f4c33b732be2fe386b7af0101a6d6789f554df03de05a694fe575b9e5" exitCode=0 Mar 21 09:00:19 crc kubenswrapper[4696]: I0321 09:00:19.121463 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ckxp8" event={"ID":"4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6","Type":"ContainerDied","Data":"d3d2352f4c33b732be2fe386b7af0101a6d6789f554df03de05a694fe575b9e5"} Mar 21 09:00:20 crc kubenswrapper[4696]: I0321 09:00:20.911861 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ckxp8" Mar 21 09:00:20 crc kubenswrapper[4696]: I0321 09:00:20.978526 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6-inventory-0\") pod \"4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6\" (UID: \"4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6\") " Mar 21 09:00:20 crc kubenswrapper[4696]: I0321 09:00:20.978627 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6-ssh-key-openstack-edpm-ipam\") pod \"4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6\" (UID: \"4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6\") " Mar 21 09:00:20 crc kubenswrapper[4696]: I0321 09:00:20.978690 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqmpj\" (UniqueName: \"kubernetes.io/projected/4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6-kube-api-access-tqmpj\") pod \"4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6\" (UID: \"4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6\") " Mar 21 09:00:20 crc kubenswrapper[4696]: I0321 09:00:20.984695 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6-kube-api-access-tqmpj" (OuterVolumeSpecName: "kube-api-access-tqmpj") pod "4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6" (UID: "4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6"). InnerVolumeSpecName "kube-api-access-tqmpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.044613 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6" (UID: "4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.054990 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6" (UID: "4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.081128 4696 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6-inventory-0\") on node \"crc\" DevicePath \"\"" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.081397 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.081478 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqmpj\" (UniqueName: \"kubernetes.io/projected/4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6-kube-api-access-tqmpj\") on node \"crc\" DevicePath \"\"" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.146063 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ckxp8" event={"ID":"4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6","Type":"ContainerDied","Data":"c6cb43a9b89a21714e3267d7ead5f94af02a4768f516cc543e83bedbd98ae7aa"} Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.146103 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6cb43a9b89a21714e3267d7ead5f94af02a4768f516cc543e83bedbd98ae7aa" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.146151 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ckxp8" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.216024 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-km5wx"] Mar 21 09:00:21 crc kubenswrapper[4696]: E0321 09:00:21.216681 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6" containerName="ssh-known-hosts-edpm-deployment" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.216748 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6" containerName="ssh-known-hosts-edpm-deployment" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.217036 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6" containerName="ssh-known-hosts-edpm-deployment" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.217939 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-km5wx" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.221224 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bqp9r" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.222000 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.222008 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.222124 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.232113 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-km5wx"] Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.287718 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-km5wx\" (UID: \"5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-km5wx" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.288575 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh77w\" (UniqueName: \"kubernetes.io/projected/5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f-kube-api-access-sh77w\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-km5wx\" (UID: \"5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-km5wx" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.288806 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-km5wx\" (UID: \"5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-km5wx" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.390474 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-km5wx\" (UID: \"5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-km5wx" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.390684 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh77w\" (UniqueName: \"kubernetes.io/projected/5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f-kube-api-access-sh77w\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-km5wx\" (UID: \"5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-km5wx" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.390734 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-km5wx\" (UID: \"5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-km5wx" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.395555 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-km5wx\" (UID: \"5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-km5wx" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.396376 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-km5wx\" (UID: \"5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-km5wx" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.408160 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh77w\" (UniqueName: \"kubernetes.io/projected/5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f-kube-api-access-sh77w\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-km5wx\" (UID: \"5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-km5wx" Mar 21 09:00:21 crc kubenswrapper[4696]: I0321 09:00:21.534088 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-km5wx" Mar 21 09:00:22 crc kubenswrapper[4696]: I0321 09:00:22.061342 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-km5wx"] Mar 21 09:00:22 crc kubenswrapper[4696]: I0321 09:00:22.156465 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-km5wx" event={"ID":"5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f","Type":"ContainerStarted","Data":"00eaa40c3aeb547519a6e1e5e60b6171ce7a67e77f7264f619e0acb9b0730e2d"} Mar 21 09:00:23 crc kubenswrapper[4696]: I0321 09:00:23.076513 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-vzfrk"] Mar 21 09:00:23 crc kubenswrapper[4696]: I0321 09:00:23.095027 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-vzfrk"] Mar 21 09:00:23 crc kubenswrapper[4696]: I0321 09:00:23.170312 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-km5wx" event={"ID":"5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f","Type":"ContainerStarted","Data":"2a19cd5a00db79b456be4c437d78087943537f9e3fc3ed892d76d52f681a0d61"} Mar 21 09:00:23 crc kubenswrapper[4696]: I0321 09:00:23.204473 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-km5wx" podStartSLOduration=1.659798726 podStartE2EDuration="2.204448935s" podCreationTimestamp="2026-03-21 09:00:21 +0000 UTC" firstStartedPulling="2026-03-21 09:00:22.063111593 +0000 UTC m=+1956.183992316" lastFinishedPulling="2026-03-21 09:00:22.607761822 +0000 UTC m=+1956.728642525" observedRunningTime="2026-03-21 09:00:23.187312226 +0000 UTC m=+1957.308192939" watchObservedRunningTime="2026-03-21 09:00:23.204448935 +0000 UTC m=+1957.325329658" Mar 21 09:00:24 crc kubenswrapper[4696]: I0321 09:00:24.545348 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="507db733-1a22-4cb0-8b84-63746b69ea4f" path="/var/lib/kubelet/pods/507db733-1a22-4cb0-8b84-63746b69ea4f/volumes" Mar 21 09:00:30 crc kubenswrapper[4696]: I0321 09:00:30.235510 4696 generic.go:334] "Generic (PLEG): container finished" podID="5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f" containerID="2a19cd5a00db79b456be4c437d78087943537f9e3fc3ed892d76d52f681a0d61" exitCode=0 Mar 21 09:00:30 crc kubenswrapper[4696]: I0321 09:00:30.235596 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-km5wx" event={"ID":"5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f","Type":"ContainerDied","Data":"2a19cd5a00db79b456be4c437d78087943537f9e3fc3ed892d76d52f681a0d61"} Mar 21 09:00:31 crc kubenswrapper[4696]: I0321 09:00:31.908803 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-km5wx" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.003502 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f-inventory\") pod \"5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f\" (UID: \"5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f\") " Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.003745 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sh77w\" (UniqueName: \"kubernetes.io/projected/5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f-kube-api-access-sh77w\") pod \"5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f\" (UID: \"5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f\") " Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.003766 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f-ssh-key-openstack-edpm-ipam\") pod \"5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f\" (UID: \"5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f\") " Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.015209 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f-kube-api-access-sh77w" (OuterVolumeSpecName: "kube-api-access-sh77w") pod "5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f" (UID: "5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f"). InnerVolumeSpecName "kube-api-access-sh77w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.031170 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f-inventory" (OuterVolumeSpecName: "inventory") pod "5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f" (UID: "5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.054662 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f" (UID: "5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.106229 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sh77w\" (UniqueName: \"kubernetes.io/projected/5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f-kube-api-access-sh77w\") on node \"crc\" DevicePath \"\"" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.106268 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.106283 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.253464 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-km5wx" event={"ID":"5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f","Type":"ContainerDied","Data":"00eaa40c3aeb547519a6e1e5e60b6171ce7a67e77f7264f619e0acb9b0730e2d"} Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.253766 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00eaa40c3aeb547519a6e1e5e60b6171ce7a67e77f7264f619e0acb9b0730e2d" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.253544 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-km5wx" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.328943 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h"] Mar 21 09:00:32 crc kubenswrapper[4696]: E0321 09:00:32.329419 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.329436 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.329638 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.330476 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.332458 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.332719 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.334548 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bqp9r" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.334550 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.355938 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h"] Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.412623 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e19a30a1-9816-4c71-9803-ed0dbbdad615-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h\" (UID: \"e19a30a1-9816-4c71-9803-ed0dbbdad615\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.412712 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e19a30a1-9816-4c71-9803-ed0dbbdad615-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h\" (UID: \"e19a30a1-9816-4c71-9803-ed0dbbdad615\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.412783 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlbt6\" (UniqueName: \"kubernetes.io/projected/e19a30a1-9816-4c71-9803-ed0dbbdad615-kube-api-access-zlbt6\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h\" (UID: \"e19a30a1-9816-4c71-9803-ed0dbbdad615\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.515170 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlbt6\" (UniqueName: \"kubernetes.io/projected/e19a30a1-9816-4c71-9803-ed0dbbdad615-kube-api-access-zlbt6\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h\" (UID: \"e19a30a1-9816-4c71-9803-ed0dbbdad615\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.515399 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e19a30a1-9816-4c71-9803-ed0dbbdad615-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h\" (UID: \"e19a30a1-9816-4c71-9803-ed0dbbdad615\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.515450 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e19a30a1-9816-4c71-9803-ed0dbbdad615-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h\" (UID: \"e19a30a1-9816-4c71-9803-ed0dbbdad615\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.519750 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e19a30a1-9816-4c71-9803-ed0dbbdad615-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h\" (UID: \"e19a30a1-9816-4c71-9803-ed0dbbdad615\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.520840 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e19a30a1-9816-4c71-9803-ed0dbbdad615-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h\" (UID: \"e19a30a1-9816-4c71-9803-ed0dbbdad615\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.531250 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlbt6\" (UniqueName: \"kubernetes.io/projected/e19a30a1-9816-4c71-9803-ed0dbbdad615-kube-api-access-zlbt6\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h\" (UID: \"e19a30a1-9816-4c71-9803-ed0dbbdad615\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h" Mar 21 09:00:32 crc kubenswrapper[4696]: I0321 09:00:32.659292 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h" Mar 21 09:00:33 crc kubenswrapper[4696]: I0321 09:00:33.187289 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h"] Mar 21 09:00:33 crc kubenswrapper[4696]: W0321 09:00:33.189971 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode19a30a1_9816_4c71_9803_ed0dbbdad615.slice/crio-b8b75f0bc530f41f261f3450259f5120603cf001c4103900b2970c5c0b4c8f67 WatchSource:0}: Error finding container b8b75f0bc530f41f261f3450259f5120603cf001c4103900b2970c5c0b4c8f67: Status 404 returned error can't find the container with id b8b75f0bc530f41f261f3450259f5120603cf001c4103900b2970c5c0b4c8f67 Mar 21 09:00:33 crc kubenswrapper[4696]: I0321 09:00:33.262930 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h" event={"ID":"e19a30a1-9816-4c71-9803-ed0dbbdad615","Type":"ContainerStarted","Data":"b8b75f0bc530f41f261f3450259f5120603cf001c4103900b2970c5c0b4c8f67"} Mar 21 09:00:34 crc kubenswrapper[4696]: I0321 09:00:34.271577 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h" event={"ID":"e19a30a1-9816-4c71-9803-ed0dbbdad615","Type":"ContainerStarted","Data":"49ca9124f0fde4bf7446905b01861e069fccd381c07659638aa7d0d1a84bc2af"} Mar 21 09:00:34 crc kubenswrapper[4696]: I0321 09:00:34.298059 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h" podStartSLOduration=1.831570686 podStartE2EDuration="2.298035053s" podCreationTimestamp="2026-03-21 09:00:32 +0000 UTC" firstStartedPulling="2026-03-21 09:00:33.191869214 +0000 UTC m=+1967.312749917" lastFinishedPulling="2026-03-21 09:00:33.658333571 +0000 UTC m=+1967.779214284" observedRunningTime="2026-03-21 09:00:34.286443686 +0000 UTC m=+1968.407324409" watchObservedRunningTime="2026-03-21 09:00:34.298035053 +0000 UTC m=+1968.418915766" Mar 21 09:00:43 crc kubenswrapper[4696]: I0321 09:00:43.348918 4696 generic.go:334] "Generic (PLEG): container finished" podID="e19a30a1-9816-4c71-9803-ed0dbbdad615" containerID="49ca9124f0fde4bf7446905b01861e069fccd381c07659638aa7d0d1a84bc2af" exitCode=0 Mar 21 09:00:43 crc kubenswrapper[4696]: I0321 09:00:43.349004 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h" event={"ID":"e19a30a1-9816-4c71-9803-ed0dbbdad615","Type":"ContainerDied","Data":"49ca9124f0fde4bf7446905b01861e069fccd381c07659638aa7d0d1a84bc2af"} Mar 21 09:00:45 crc kubenswrapper[4696]: I0321 09:00:45.831776 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h" Mar 21 09:00:45 crc kubenswrapper[4696]: I0321 09:00:45.898045 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e19a30a1-9816-4c71-9803-ed0dbbdad615-ssh-key-openstack-edpm-ipam\") pod \"e19a30a1-9816-4c71-9803-ed0dbbdad615\" (UID: \"e19a30a1-9816-4c71-9803-ed0dbbdad615\") " Mar 21 09:00:45 crc kubenswrapper[4696]: I0321 09:00:45.898158 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e19a30a1-9816-4c71-9803-ed0dbbdad615-inventory\") pod \"e19a30a1-9816-4c71-9803-ed0dbbdad615\" (UID: \"e19a30a1-9816-4c71-9803-ed0dbbdad615\") " Mar 21 09:00:45 crc kubenswrapper[4696]: I0321 09:00:45.898412 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlbt6\" (UniqueName: \"kubernetes.io/projected/e19a30a1-9816-4c71-9803-ed0dbbdad615-kube-api-access-zlbt6\") pod \"e19a30a1-9816-4c71-9803-ed0dbbdad615\" (UID: \"e19a30a1-9816-4c71-9803-ed0dbbdad615\") " Mar 21 09:00:45 crc kubenswrapper[4696]: I0321 09:00:45.944188 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e19a30a1-9816-4c71-9803-ed0dbbdad615-kube-api-access-zlbt6" (OuterVolumeSpecName: "kube-api-access-zlbt6") pod "e19a30a1-9816-4c71-9803-ed0dbbdad615" (UID: "e19a30a1-9816-4c71-9803-ed0dbbdad615"). InnerVolumeSpecName "kube-api-access-zlbt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:00:46 crc kubenswrapper[4696]: I0321 09:00:46.002386 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlbt6\" (UniqueName: \"kubernetes.io/projected/e19a30a1-9816-4c71-9803-ed0dbbdad615-kube-api-access-zlbt6\") on node \"crc\" DevicePath \"\"" Mar 21 09:00:46 crc kubenswrapper[4696]: I0321 09:00:46.023194 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e19a30a1-9816-4c71-9803-ed0dbbdad615-inventory" (OuterVolumeSpecName: "inventory") pod "e19a30a1-9816-4c71-9803-ed0dbbdad615" (UID: "e19a30a1-9816-4c71-9803-ed0dbbdad615"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:00:46 crc kubenswrapper[4696]: I0321 09:00:46.023583 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e19a30a1-9816-4c71-9803-ed0dbbdad615-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e19a30a1-9816-4c71-9803-ed0dbbdad615" (UID: "e19a30a1-9816-4c71-9803-ed0dbbdad615"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:00:46 crc kubenswrapper[4696]: I0321 09:00:46.104230 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e19a30a1-9816-4c71-9803-ed0dbbdad615-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 09:00:46 crc kubenswrapper[4696]: I0321 09:00:46.104276 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e19a30a1-9816-4c71-9803-ed0dbbdad615-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 09:00:46 crc kubenswrapper[4696]: I0321 09:00:46.481529 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h" event={"ID":"e19a30a1-9816-4c71-9803-ed0dbbdad615","Type":"ContainerDied","Data":"b8b75f0bc530f41f261f3450259f5120603cf001c4103900b2970c5c0b4c8f67"} Mar 21 09:00:46 crc kubenswrapper[4696]: I0321 09:00:46.481580 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h" Mar 21 09:00:46 crc kubenswrapper[4696]: I0321 09:00:46.481586 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8b75f0bc530f41f261f3450259f5120603cf001c4103900b2970c5c0b4c8f67" Mar 21 09:00:46 crc kubenswrapper[4696]: I0321 09:00:46.934305 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2"] Mar 21 09:00:46 crc kubenswrapper[4696]: E0321 09:00:46.935517 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e19a30a1-9816-4c71-9803-ed0dbbdad615" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 21 09:00:46 crc kubenswrapper[4696]: I0321 09:00:46.935533 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e19a30a1-9816-4c71-9803-ed0dbbdad615" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 21 09:00:46 crc kubenswrapper[4696]: I0321 09:00:46.935737 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="e19a30a1-9816-4c71-9803-ed0dbbdad615" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 21 09:00:46 crc kubenswrapper[4696]: I0321 09:00:46.936626 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:46 crc kubenswrapper[4696]: I0321 09:00:46.941435 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Mar 21 09:00:46 crc kubenswrapper[4696]: I0321 09:00:46.941570 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Mar 21 09:00:46 crc kubenswrapper[4696]: I0321 09:00:46.941900 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 09:00:46 crc kubenswrapper[4696]: I0321 09:00:46.942025 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 09:00:46 crc kubenswrapper[4696]: I0321 09:00:46.942045 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bqp9r" Mar 21 09:00:46 crc kubenswrapper[4696]: I0321 09:00:46.942121 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Mar 21 09:00:46 crc kubenswrapper[4696]: I0321 09:00:46.942542 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Mar 21 09:00:46 crc kubenswrapper[4696]: I0321 09:00:46.947060 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 09:00:46 crc kubenswrapper[4696]: I0321 09:00:46.953515 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2"] Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.022003 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.022064 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.022118 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.022151 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.022194 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.022251 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjg75\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-kube-api-access-pjg75\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.022302 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.022370 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.022476 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.022538 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.022669 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.022746 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.022768 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.022788 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.124727 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.124783 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.124821 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.124865 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.124886 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.125515 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.125562 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.125585 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.125614 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.125638 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.125934 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.125972 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjg75\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-kube-api-access-pjg75\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.126001 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.126037 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.130040 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.130488 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.130625 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.131446 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.132408 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.132701 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.133777 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.134235 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.134695 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.136455 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.141466 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.142295 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.145199 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjg75\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-kube-api-access-pjg75\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.145800 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.258120 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:00:47 crc kubenswrapper[4696]: I0321 09:00:47.845179 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2"] Mar 21 09:00:48 crc kubenswrapper[4696]: I0321 09:00:48.503260 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" event={"ID":"053326a0-54a2-4838-b024-4650e216f804","Type":"ContainerStarted","Data":"8c2e515afe96a1566a3c519a9752a624d6bdc701c3550770e74e343faee6dbce"} Mar 21 09:00:49 crc kubenswrapper[4696]: I0321 09:00:49.513204 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" event={"ID":"053326a0-54a2-4838-b024-4650e216f804","Type":"ContainerStarted","Data":"8de22c83333bdc05dac80fc790c24072aba794abcabc8cdf89d297b7c1e50da5"} Mar 21 09:00:49 crc kubenswrapper[4696]: I0321 09:00:49.536055 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" podStartSLOduration=3.026944328 podStartE2EDuration="3.536036952s" podCreationTimestamp="2026-03-21 09:00:46 +0000 UTC" firstStartedPulling="2026-03-21 09:00:47.857481204 +0000 UTC m=+1981.978361917" lastFinishedPulling="2026-03-21 09:00:48.366573828 +0000 UTC m=+1982.487454541" observedRunningTime="2026-03-21 09:00:49.531202629 +0000 UTC m=+1983.652083372" watchObservedRunningTime="2026-03-21 09:00:49.536036952 +0000 UTC m=+1983.656917665" Mar 21 09:01:00 crc kubenswrapper[4696]: I0321 09:01:00.159667 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29568061-9qzjv"] Mar 21 09:01:00 crc kubenswrapper[4696]: I0321 09:01:00.161618 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29568061-9qzjv" Mar 21 09:01:00 crc kubenswrapper[4696]: I0321 09:01:00.186743 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29568061-9qzjv"] Mar 21 09:01:00 crc kubenswrapper[4696]: I0321 09:01:00.217730 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93e69ad1-eb81-4a1a-ae60-6feff0891ced-fernet-keys\") pod \"keystone-cron-29568061-9qzjv\" (UID: \"93e69ad1-eb81-4a1a-ae60-6feff0891ced\") " pod="openstack/keystone-cron-29568061-9qzjv" Mar 21 09:01:00 crc kubenswrapper[4696]: I0321 09:01:00.217798 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93e69ad1-eb81-4a1a-ae60-6feff0891ced-config-data\") pod \"keystone-cron-29568061-9qzjv\" (UID: \"93e69ad1-eb81-4a1a-ae60-6feff0891ced\") " pod="openstack/keystone-cron-29568061-9qzjv" Mar 21 09:01:00 crc kubenswrapper[4696]: I0321 09:01:00.217845 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq4zb\" (UniqueName: \"kubernetes.io/projected/93e69ad1-eb81-4a1a-ae60-6feff0891ced-kube-api-access-hq4zb\") pod \"keystone-cron-29568061-9qzjv\" (UID: \"93e69ad1-eb81-4a1a-ae60-6feff0891ced\") " pod="openstack/keystone-cron-29568061-9qzjv" Mar 21 09:01:00 crc kubenswrapper[4696]: I0321 09:01:00.218225 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e69ad1-eb81-4a1a-ae60-6feff0891ced-combined-ca-bundle\") pod \"keystone-cron-29568061-9qzjv\" (UID: \"93e69ad1-eb81-4a1a-ae60-6feff0891ced\") " pod="openstack/keystone-cron-29568061-9qzjv" Mar 21 09:01:00 crc kubenswrapper[4696]: I0321 09:01:00.320085 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93e69ad1-eb81-4a1a-ae60-6feff0891ced-fernet-keys\") pod \"keystone-cron-29568061-9qzjv\" (UID: \"93e69ad1-eb81-4a1a-ae60-6feff0891ced\") " pod="openstack/keystone-cron-29568061-9qzjv" Mar 21 09:01:00 crc kubenswrapper[4696]: I0321 09:01:00.320132 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93e69ad1-eb81-4a1a-ae60-6feff0891ced-config-data\") pod \"keystone-cron-29568061-9qzjv\" (UID: \"93e69ad1-eb81-4a1a-ae60-6feff0891ced\") " pod="openstack/keystone-cron-29568061-9qzjv" Mar 21 09:01:00 crc kubenswrapper[4696]: I0321 09:01:00.320151 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq4zb\" (UniqueName: \"kubernetes.io/projected/93e69ad1-eb81-4a1a-ae60-6feff0891ced-kube-api-access-hq4zb\") pod \"keystone-cron-29568061-9qzjv\" (UID: \"93e69ad1-eb81-4a1a-ae60-6feff0891ced\") " pod="openstack/keystone-cron-29568061-9qzjv" Mar 21 09:01:00 crc kubenswrapper[4696]: I0321 09:01:00.320210 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e69ad1-eb81-4a1a-ae60-6feff0891ced-combined-ca-bundle\") pod \"keystone-cron-29568061-9qzjv\" (UID: \"93e69ad1-eb81-4a1a-ae60-6feff0891ced\") " pod="openstack/keystone-cron-29568061-9qzjv" Mar 21 09:01:00 crc kubenswrapper[4696]: I0321 09:01:00.330073 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93e69ad1-eb81-4a1a-ae60-6feff0891ced-config-data\") pod \"keystone-cron-29568061-9qzjv\" (UID: \"93e69ad1-eb81-4a1a-ae60-6feff0891ced\") " pod="openstack/keystone-cron-29568061-9qzjv" Mar 21 09:01:00 crc kubenswrapper[4696]: I0321 09:01:00.330797 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93e69ad1-eb81-4a1a-ae60-6feff0891ced-fernet-keys\") pod \"keystone-cron-29568061-9qzjv\" (UID: \"93e69ad1-eb81-4a1a-ae60-6feff0891ced\") " pod="openstack/keystone-cron-29568061-9qzjv" Mar 21 09:01:00 crc kubenswrapper[4696]: I0321 09:01:00.331622 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e69ad1-eb81-4a1a-ae60-6feff0891ced-combined-ca-bundle\") pod \"keystone-cron-29568061-9qzjv\" (UID: \"93e69ad1-eb81-4a1a-ae60-6feff0891ced\") " pod="openstack/keystone-cron-29568061-9qzjv" Mar 21 09:01:00 crc kubenswrapper[4696]: I0321 09:01:00.339587 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq4zb\" (UniqueName: \"kubernetes.io/projected/93e69ad1-eb81-4a1a-ae60-6feff0891ced-kube-api-access-hq4zb\") pod \"keystone-cron-29568061-9qzjv\" (UID: \"93e69ad1-eb81-4a1a-ae60-6feff0891ced\") " pod="openstack/keystone-cron-29568061-9qzjv" Mar 21 09:01:00 crc kubenswrapper[4696]: I0321 09:01:00.480159 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29568061-9qzjv" Mar 21 09:01:00 crc kubenswrapper[4696]: I0321 09:01:00.958646 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29568061-9qzjv"] Mar 21 09:01:00 crc kubenswrapper[4696]: W0321 09:01:00.978733 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93e69ad1_eb81_4a1a_ae60_6feff0891ced.slice/crio-509797a62f5b4edb17500d0cfb7d1fd809b724068a16ca7d38c77ebac2798582 WatchSource:0}: Error finding container 509797a62f5b4edb17500d0cfb7d1fd809b724068a16ca7d38c77ebac2798582: Status 404 returned error can't find the container with id 509797a62f5b4edb17500d0cfb7d1fd809b724068a16ca7d38c77ebac2798582 Mar 21 09:01:01 crc kubenswrapper[4696]: I0321 09:01:01.148558 4696 scope.go:117] "RemoveContainer" containerID="9886082c780636b6eb0d3b13b235a0763fe4ce7830032888f8c170f2e7893b1c" Mar 21 09:01:01 crc kubenswrapper[4696]: I0321 09:01:01.222123 4696 scope.go:117] "RemoveContainer" containerID="26fd12ef9eb9bd83c4e74724cce711f57f7190be8e20833ce132ed584b0c3116" Mar 21 09:01:01 crc kubenswrapper[4696]: I0321 09:01:01.245308 4696 scope.go:117] "RemoveContainer" containerID="5c2f857bc27ae26eded78f4c0167984e206a5936f4a62a234f63d4edb7f3b89f" Mar 21 09:01:01 crc kubenswrapper[4696]: I0321 09:01:01.611355 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29568061-9qzjv" event={"ID":"93e69ad1-eb81-4a1a-ae60-6feff0891ced","Type":"ContainerStarted","Data":"707eb2e0e76b45daae9ad99e4175499dc16bca994cae10f2c23e1536eeb3e150"} Mar 21 09:01:01 crc kubenswrapper[4696]: I0321 09:01:01.611687 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29568061-9qzjv" event={"ID":"93e69ad1-eb81-4a1a-ae60-6feff0891ced","Type":"ContainerStarted","Data":"509797a62f5b4edb17500d0cfb7d1fd809b724068a16ca7d38c77ebac2798582"} Mar 21 09:01:01 crc kubenswrapper[4696]: I0321 09:01:01.628669 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29568061-9qzjv" podStartSLOduration=1.628650543 podStartE2EDuration="1.628650543s" podCreationTimestamp="2026-03-21 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 09:01:01.627732179 +0000 UTC m=+1995.748612892" watchObservedRunningTime="2026-03-21 09:01:01.628650543 +0000 UTC m=+1995.749531256" Mar 21 09:01:04 crc kubenswrapper[4696]: I0321 09:01:04.640124 4696 generic.go:334] "Generic (PLEG): container finished" podID="93e69ad1-eb81-4a1a-ae60-6feff0891ced" containerID="707eb2e0e76b45daae9ad99e4175499dc16bca994cae10f2c23e1536eeb3e150" exitCode=0 Mar 21 09:01:04 crc kubenswrapper[4696]: I0321 09:01:04.640213 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29568061-9qzjv" event={"ID":"93e69ad1-eb81-4a1a-ae60-6feff0891ced","Type":"ContainerDied","Data":"707eb2e0e76b45daae9ad99e4175499dc16bca994cae10f2c23e1536eeb3e150"} Mar 21 09:01:06 crc kubenswrapper[4696]: I0321 09:01:06.102620 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29568061-9qzjv" Mar 21 09:01:06 crc kubenswrapper[4696]: I0321 09:01:06.240208 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e69ad1-eb81-4a1a-ae60-6feff0891ced-combined-ca-bundle\") pod \"93e69ad1-eb81-4a1a-ae60-6feff0891ced\" (UID: \"93e69ad1-eb81-4a1a-ae60-6feff0891ced\") " Mar 21 09:01:06 crc kubenswrapper[4696]: I0321 09:01:06.240474 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hq4zb\" (UniqueName: \"kubernetes.io/projected/93e69ad1-eb81-4a1a-ae60-6feff0891ced-kube-api-access-hq4zb\") pod \"93e69ad1-eb81-4a1a-ae60-6feff0891ced\" (UID: \"93e69ad1-eb81-4a1a-ae60-6feff0891ced\") " Mar 21 09:01:06 crc kubenswrapper[4696]: I0321 09:01:06.240759 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93e69ad1-eb81-4a1a-ae60-6feff0891ced-fernet-keys\") pod \"93e69ad1-eb81-4a1a-ae60-6feff0891ced\" (UID: \"93e69ad1-eb81-4a1a-ae60-6feff0891ced\") " Mar 21 09:01:06 crc kubenswrapper[4696]: I0321 09:01:06.240844 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93e69ad1-eb81-4a1a-ae60-6feff0891ced-config-data\") pod \"93e69ad1-eb81-4a1a-ae60-6feff0891ced\" (UID: \"93e69ad1-eb81-4a1a-ae60-6feff0891ced\") " Mar 21 09:01:06 crc kubenswrapper[4696]: I0321 09:01:06.248002 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93e69ad1-eb81-4a1a-ae60-6feff0891ced-kube-api-access-hq4zb" (OuterVolumeSpecName: "kube-api-access-hq4zb") pod "93e69ad1-eb81-4a1a-ae60-6feff0891ced" (UID: "93e69ad1-eb81-4a1a-ae60-6feff0891ced"). InnerVolumeSpecName "kube-api-access-hq4zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:01:06 crc kubenswrapper[4696]: I0321 09:01:06.248326 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e69ad1-eb81-4a1a-ae60-6feff0891ced-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "93e69ad1-eb81-4a1a-ae60-6feff0891ced" (UID: "93e69ad1-eb81-4a1a-ae60-6feff0891ced"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:01:06 crc kubenswrapper[4696]: I0321 09:01:06.270511 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e69ad1-eb81-4a1a-ae60-6feff0891ced-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93e69ad1-eb81-4a1a-ae60-6feff0891ced" (UID: "93e69ad1-eb81-4a1a-ae60-6feff0891ced"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:01:06 crc kubenswrapper[4696]: I0321 09:01:06.294071 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e69ad1-eb81-4a1a-ae60-6feff0891ced-config-data" (OuterVolumeSpecName: "config-data") pod "93e69ad1-eb81-4a1a-ae60-6feff0891ced" (UID: "93e69ad1-eb81-4a1a-ae60-6feff0891ced"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:01:06 crc kubenswrapper[4696]: I0321 09:01:06.344123 4696 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e69ad1-eb81-4a1a-ae60-6feff0891ced-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 09:01:06 crc kubenswrapper[4696]: I0321 09:01:06.344183 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hq4zb\" (UniqueName: \"kubernetes.io/projected/93e69ad1-eb81-4a1a-ae60-6feff0891ced-kube-api-access-hq4zb\") on node \"crc\" DevicePath \"\"" Mar 21 09:01:06 crc kubenswrapper[4696]: I0321 09:01:06.344202 4696 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93e69ad1-eb81-4a1a-ae60-6feff0891ced-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 21 09:01:06 crc kubenswrapper[4696]: I0321 09:01:06.344213 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93e69ad1-eb81-4a1a-ae60-6feff0891ced-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 09:01:06 crc kubenswrapper[4696]: I0321 09:01:06.657958 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29568061-9qzjv" event={"ID":"93e69ad1-eb81-4a1a-ae60-6feff0891ced","Type":"ContainerDied","Data":"509797a62f5b4edb17500d0cfb7d1fd809b724068a16ca7d38c77ebac2798582"} Mar 21 09:01:06 crc kubenswrapper[4696]: I0321 09:01:06.658299 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="509797a62f5b4edb17500d0cfb7d1fd809b724068a16ca7d38c77ebac2798582" Mar 21 09:01:06 crc kubenswrapper[4696]: I0321 09:01:06.658073 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29568061-9qzjv" Mar 21 09:01:23 crc kubenswrapper[4696]: I0321 09:01:23.806140 4696 generic.go:334] "Generic (PLEG): container finished" podID="053326a0-54a2-4838-b024-4650e216f804" containerID="8de22c83333bdc05dac80fc790c24072aba794abcabc8cdf89d297b7c1e50da5" exitCode=0 Mar 21 09:01:23 crc kubenswrapper[4696]: I0321 09:01:23.806233 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" event={"ID":"053326a0-54a2-4838-b024-4650e216f804","Type":"ContainerDied","Data":"8de22c83333bdc05dac80fc790c24072aba794abcabc8cdf89d297b7c1e50da5"} Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.311671 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.425491 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-ovn-default-certs-0\") pod \"053326a0-54a2-4838-b024-4650e216f804\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.425616 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-bootstrap-combined-ca-bundle\") pod \"053326a0-54a2-4838-b024-4650e216f804\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.425658 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-neutron-metadata-combined-ca-bundle\") pod \"053326a0-54a2-4838-b024-4650e216f804\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.425811 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjg75\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-kube-api-access-pjg75\") pod \"053326a0-54a2-4838-b024-4650e216f804\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.425932 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-inventory\") pod \"053326a0-54a2-4838-b024-4650e216f804\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.425992 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-repo-setup-combined-ca-bundle\") pod \"053326a0-54a2-4838-b024-4650e216f804\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.426039 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-ovn-combined-ca-bundle\") pod \"053326a0-54a2-4838-b024-4650e216f804\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.426116 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-libvirt-combined-ca-bundle\") pod \"053326a0-54a2-4838-b024-4650e216f804\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.426158 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"053326a0-54a2-4838-b024-4650e216f804\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.426249 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"053326a0-54a2-4838-b024-4650e216f804\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.426298 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-nova-combined-ca-bundle\") pod \"053326a0-54a2-4838-b024-4650e216f804\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.426343 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-telemetry-combined-ca-bundle\") pod \"053326a0-54a2-4838-b024-4650e216f804\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.426377 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-ssh-key-openstack-edpm-ipam\") pod \"053326a0-54a2-4838-b024-4650e216f804\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.426483 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"053326a0-54a2-4838-b024-4650e216f804\" (UID: \"053326a0-54a2-4838-b024-4650e216f804\") " Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.437114 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "053326a0-54a2-4838-b024-4650e216f804" (UID: "053326a0-54a2-4838-b024-4650e216f804"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.437166 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "053326a0-54a2-4838-b024-4650e216f804" (UID: "053326a0-54a2-4838-b024-4650e216f804"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.437239 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "053326a0-54a2-4838-b024-4650e216f804" (UID: "053326a0-54a2-4838-b024-4650e216f804"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.437263 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "053326a0-54a2-4838-b024-4650e216f804" (UID: "053326a0-54a2-4838-b024-4650e216f804"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.437328 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "053326a0-54a2-4838-b024-4650e216f804" (UID: "053326a0-54a2-4838-b024-4650e216f804"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.437422 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-kube-api-access-pjg75" (OuterVolumeSpecName: "kube-api-access-pjg75") pod "053326a0-54a2-4838-b024-4650e216f804" (UID: "053326a0-54a2-4838-b024-4650e216f804"). InnerVolumeSpecName "kube-api-access-pjg75". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.438020 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "053326a0-54a2-4838-b024-4650e216f804" (UID: "053326a0-54a2-4838-b024-4650e216f804"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.438806 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "053326a0-54a2-4838-b024-4650e216f804" (UID: "053326a0-54a2-4838-b024-4650e216f804"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.438910 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "053326a0-54a2-4838-b024-4650e216f804" (UID: "053326a0-54a2-4838-b024-4650e216f804"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.439518 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "053326a0-54a2-4838-b024-4650e216f804" (UID: "053326a0-54a2-4838-b024-4650e216f804"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.443161 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "053326a0-54a2-4838-b024-4650e216f804" (UID: "053326a0-54a2-4838-b024-4650e216f804"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.453124 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "053326a0-54a2-4838-b024-4650e216f804" (UID: "053326a0-54a2-4838-b024-4650e216f804"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.467618 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "053326a0-54a2-4838-b024-4650e216f804" (UID: "053326a0-54a2-4838-b024-4650e216f804"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.491405 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-inventory" (OuterVolumeSpecName: "inventory") pod "053326a0-54a2-4838-b024-4650e216f804" (UID: "053326a0-54a2-4838-b024-4650e216f804"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.529535 4696 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.529589 4696 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.529605 4696 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.529623 4696 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.529638 4696 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.529653 4696 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.529669 4696 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.529685 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.529699 4696 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.529713 4696 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.529727 4696 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.529742 4696 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.529760 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjg75\" (UniqueName: \"kubernetes.io/projected/053326a0-54a2-4838-b024-4650e216f804-kube-api-access-pjg75\") on node \"crc\" DevicePath \"\"" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.529775 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/053326a0-54a2-4838-b024-4650e216f804-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.829969 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" event={"ID":"053326a0-54a2-4838-b024-4650e216f804","Type":"ContainerDied","Data":"8c2e515afe96a1566a3c519a9752a624d6bdc701c3550770e74e343faee6dbce"} Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.830011 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c2e515afe96a1566a3c519a9752a624d6bdc701c3550770e74e343faee6dbce" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.830020 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.921979 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz"] Mar 21 09:01:25 crc kubenswrapper[4696]: E0321 09:01:25.923534 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e69ad1-eb81-4a1a-ae60-6feff0891ced" containerName="keystone-cron" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.923558 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e69ad1-eb81-4a1a-ae60-6feff0891ced" containerName="keystone-cron" Mar 21 09:01:25 crc kubenswrapper[4696]: E0321 09:01:25.923622 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="053326a0-54a2-4838-b024-4650e216f804" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.923633 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="053326a0-54a2-4838-b024-4650e216f804" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.923885 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="053326a0-54a2-4838-b024-4650e216f804" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.923926 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e69ad1-eb81-4a1a-ae60-6feff0891ced" containerName="keystone-cron" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.924863 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.927946 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.928164 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.928362 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.928493 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.929447 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bqp9r" Mar 21 09:01:25 crc kubenswrapper[4696]: I0321 09:01:25.940797 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz"] Mar 21 09:01:26 crc kubenswrapper[4696]: I0321 09:01:26.042347 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/2c75b9c9-8d94-4541-8858-0c56d8ecb400-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5hcz\" (UID: \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" Mar 21 09:01:26 crc kubenswrapper[4696]: I0321 09:01:26.043074 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5l5p\" (UniqueName: \"kubernetes.io/projected/2c75b9c9-8d94-4541-8858-0c56d8ecb400-kube-api-access-x5l5p\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5hcz\" (UID: \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" Mar 21 09:01:26 crc kubenswrapper[4696]: I0321 09:01:26.043477 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c75b9c9-8d94-4541-8858-0c56d8ecb400-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5hcz\" (UID: \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" Mar 21 09:01:26 crc kubenswrapper[4696]: I0321 09:01:26.043528 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2c75b9c9-8d94-4541-8858-0c56d8ecb400-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5hcz\" (UID: \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" Mar 21 09:01:26 crc kubenswrapper[4696]: I0321 09:01:26.043652 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c75b9c9-8d94-4541-8858-0c56d8ecb400-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5hcz\" (UID: \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" Mar 21 09:01:26 crc kubenswrapper[4696]: I0321 09:01:26.145756 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c75b9c9-8d94-4541-8858-0c56d8ecb400-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5hcz\" (UID: \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" Mar 21 09:01:26 crc kubenswrapper[4696]: I0321 09:01:26.145808 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2c75b9c9-8d94-4541-8858-0c56d8ecb400-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5hcz\" (UID: \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" Mar 21 09:01:26 crc kubenswrapper[4696]: I0321 09:01:26.145866 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c75b9c9-8d94-4541-8858-0c56d8ecb400-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5hcz\" (UID: \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" Mar 21 09:01:26 crc kubenswrapper[4696]: I0321 09:01:26.145894 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/2c75b9c9-8d94-4541-8858-0c56d8ecb400-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5hcz\" (UID: \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" Mar 21 09:01:26 crc kubenswrapper[4696]: I0321 09:01:26.145938 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5l5p\" (UniqueName: \"kubernetes.io/projected/2c75b9c9-8d94-4541-8858-0c56d8ecb400-kube-api-access-x5l5p\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5hcz\" (UID: \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" Mar 21 09:01:26 crc kubenswrapper[4696]: I0321 09:01:26.147016 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/2c75b9c9-8d94-4541-8858-0c56d8ecb400-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5hcz\" (UID: \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" Mar 21 09:01:26 crc kubenswrapper[4696]: I0321 09:01:26.149636 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c75b9c9-8d94-4541-8858-0c56d8ecb400-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5hcz\" (UID: \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" Mar 21 09:01:26 crc kubenswrapper[4696]: I0321 09:01:26.150107 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c75b9c9-8d94-4541-8858-0c56d8ecb400-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5hcz\" (UID: \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" Mar 21 09:01:26 crc kubenswrapper[4696]: I0321 09:01:26.150296 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2c75b9c9-8d94-4541-8858-0c56d8ecb400-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5hcz\" (UID: \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" Mar 21 09:01:26 crc kubenswrapper[4696]: I0321 09:01:26.164134 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5l5p\" (UniqueName: \"kubernetes.io/projected/2c75b9c9-8d94-4541-8858-0c56d8ecb400-kube-api-access-x5l5p\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-v5hcz\" (UID: \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" Mar 21 09:01:26 crc kubenswrapper[4696]: I0321 09:01:26.291380 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" Mar 21 09:01:26 crc kubenswrapper[4696]: I0321 09:01:26.919806 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz"] Mar 21 09:01:27 crc kubenswrapper[4696]: I0321 09:01:27.869996 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" event={"ID":"2c75b9c9-8d94-4541-8858-0c56d8ecb400","Type":"ContainerStarted","Data":"b0d8f7d4953b05bf4a1decfa62e02d71ec519caf4b664217a87ce6de7624b690"} Mar 21 09:01:27 crc kubenswrapper[4696]: I0321 09:01:27.870376 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" event={"ID":"2c75b9c9-8d94-4541-8858-0c56d8ecb400","Type":"ContainerStarted","Data":"9d124e9e52a2750c16d6ab29d044bb7cac8d0e4958f30cd7fa6a182b27844272"} Mar 21 09:01:27 crc kubenswrapper[4696]: I0321 09:01:27.916509 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" podStartSLOduration=2.503460637 podStartE2EDuration="2.9164645s" podCreationTimestamp="2026-03-21 09:01:25 +0000 UTC" firstStartedPulling="2026-03-21 09:01:26.923344848 +0000 UTC m=+2021.044225561" lastFinishedPulling="2026-03-21 09:01:27.336348711 +0000 UTC m=+2021.457229424" observedRunningTime="2026-03-21 09:01:27.89819642 +0000 UTC m=+2022.019077143" watchObservedRunningTime="2026-03-21 09:01:27.9164645 +0000 UTC m=+2022.037345223" Mar 21 09:01:48 crc kubenswrapper[4696]: I0321 09:01:48.046858 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-db-sync-fq9xb"] Mar 21 09:01:48 crc kubenswrapper[4696]: I0321 09:01:48.056981 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-db-sync-fq9xb"] Mar 21 09:01:48 crc kubenswrapper[4696]: I0321 09:01:48.547006 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f8a2d50-6415-4d3a-8cb4-bbbfce785842" path="/var/lib/kubelet/pods/1f8a2d50-6415-4d3a-8cb4-bbbfce785842/volumes" Mar 21 09:01:54 crc kubenswrapper[4696]: I0321 09:01:54.037658 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-storageinit-8r5zz"] Mar 21 09:01:54 crc kubenswrapper[4696]: I0321 09:01:54.048028 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-storageinit-8r5zz"] Mar 21 09:01:54 crc kubenswrapper[4696]: I0321 09:01:54.562940 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e75fe987-8f3c-4f76-9546-c6ab537af3ce" path="/var/lib/kubelet/pods/e75fe987-8f3c-4f76-9546-c6ab537af3ce/volumes" Mar 21 09:02:00 crc kubenswrapper[4696]: I0321 09:02:00.144965 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568062-twbpv"] Mar 21 09:02:00 crc kubenswrapper[4696]: I0321 09:02:00.147067 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568062-twbpv" Mar 21 09:02:00 crc kubenswrapper[4696]: I0321 09:02:00.149897 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:02:00 crc kubenswrapper[4696]: I0321 09:02:00.150557 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:02:00 crc kubenswrapper[4696]: I0321 09:02:00.150781 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:02:00 crc kubenswrapper[4696]: I0321 09:02:00.155694 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568062-twbpv"] Mar 21 09:02:00 crc kubenswrapper[4696]: I0321 09:02:00.264375 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kblzb\" (UniqueName: \"kubernetes.io/projected/d1d1e242-e372-472f-a67c-709b9d94dfe4-kube-api-access-kblzb\") pod \"auto-csr-approver-29568062-twbpv\" (UID: \"d1d1e242-e372-472f-a67c-709b9d94dfe4\") " pod="openshift-infra/auto-csr-approver-29568062-twbpv" Mar 21 09:02:00 crc kubenswrapper[4696]: I0321 09:02:00.384275 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kblzb\" (UniqueName: \"kubernetes.io/projected/d1d1e242-e372-472f-a67c-709b9d94dfe4-kube-api-access-kblzb\") pod \"auto-csr-approver-29568062-twbpv\" (UID: \"d1d1e242-e372-472f-a67c-709b9d94dfe4\") " pod="openshift-infra/auto-csr-approver-29568062-twbpv" Mar 21 09:02:00 crc kubenswrapper[4696]: I0321 09:02:00.403922 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kblzb\" (UniqueName: \"kubernetes.io/projected/d1d1e242-e372-472f-a67c-709b9d94dfe4-kube-api-access-kblzb\") pod \"auto-csr-approver-29568062-twbpv\" (UID: \"d1d1e242-e372-472f-a67c-709b9d94dfe4\") " pod="openshift-infra/auto-csr-approver-29568062-twbpv" Mar 21 09:02:00 crc kubenswrapper[4696]: I0321 09:02:00.469611 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568062-twbpv" Mar 21 09:02:00 crc kubenswrapper[4696]: I0321 09:02:00.901072 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568062-twbpv"] Mar 21 09:02:01 crc kubenswrapper[4696]: I0321 09:02:01.177180 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568062-twbpv" event={"ID":"d1d1e242-e372-472f-a67c-709b9d94dfe4","Type":"ContainerStarted","Data":"fd24dcecbb313385171d86ff3a2f310757d5b38b433a4c84d43d2fef021fbfbb"} Mar 21 09:02:01 crc kubenswrapper[4696]: I0321 09:02:01.359027 4696 scope.go:117] "RemoveContainer" containerID="f88fc9473fb238eb74c3a9d5c3c6794ca0df87320ffa9a7de517cdfb9c6f0d5b" Mar 21 09:02:01 crc kubenswrapper[4696]: I0321 09:02:01.401350 4696 scope.go:117] "RemoveContainer" containerID="5137937cfb80b886d262e80869076dfafade90bbe4c64c139378725b1d8e437d" Mar 21 09:02:02 crc kubenswrapper[4696]: I0321 09:02:02.187713 4696 generic.go:334] "Generic (PLEG): container finished" podID="d1d1e242-e372-472f-a67c-709b9d94dfe4" containerID="630731d5ea50a7b392f5de8cf9f8fca62ee44b1810f5a04d5a8f47cef8109dbb" exitCode=0 Mar 21 09:02:02 crc kubenswrapper[4696]: I0321 09:02:02.187767 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568062-twbpv" event={"ID":"d1d1e242-e372-472f-a67c-709b9d94dfe4","Type":"ContainerDied","Data":"630731d5ea50a7b392f5de8cf9f8fca62ee44b1810f5a04d5a8f47cef8109dbb"} Mar 21 09:02:03 crc kubenswrapper[4696]: I0321 09:02:03.592295 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568062-twbpv" Mar 21 09:02:03 crc kubenswrapper[4696]: I0321 09:02:03.663804 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kblzb\" (UniqueName: \"kubernetes.io/projected/d1d1e242-e372-472f-a67c-709b9d94dfe4-kube-api-access-kblzb\") pod \"d1d1e242-e372-472f-a67c-709b9d94dfe4\" (UID: \"d1d1e242-e372-472f-a67c-709b9d94dfe4\") " Mar 21 09:02:03 crc kubenswrapper[4696]: I0321 09:02:03.675036 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1d1e242-e372-472f-a67c-709b9d94dfe4-kube-api-access-kblzb" (OuterVolumeSpecName: "kube-api-access-kblzb") pod "d1d1e242-e372-472f-a67c-709b9d94dfe4" (UID: "d1d1e242-e372-472f-a67c-709b9d94dfe4"). InnerVolumeSpecName "kube-api-access-kblzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:02:03 crc kubenswrapper[4696]: I0321 09:02:03.767986 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kblzb\" (UniqueName: \"kubernetes.io/projected/d1d1e242-e372-472f-a67c-709b9d94dfe4-kube-api-access-kblzb\") on node \"crc\" DevicePath \"\"" Mar 21 09:02:04 crc kubenswrapper[4696]: I0321 09:02:04.207688 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568062-twbpv" event={"ID":"d1d1e242-e372-472f-a67c-709b9d94dfe4","Type":"ContainerDied","Data":"fd24dcecbb313385171d86ff3a2f310757d5b38b433a4c84d43d2fef021fbfbb"} Mar 21 09:02:04 crc kubenswrapper[4696]: I0321 09:02:04.208007 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd24dcecbb313385171d86ff3a2f310757d5b38b433a4c84d43d2fef021fbfbb" Mar 21 09:02:04 crc kubenswrapper[4696]: I0321 09:02:04.207725 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568062-twbpv" Mar 21 09:02:04 crc kubenswrapper[4696]: I0321 09:02:04.643143 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568056-sgpwk"] Mar 21 09:02:04 crc kubenswrapper[4696]: I0321 09:02:04.651236 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568056-sgpwk"] Mar 21 09:02:06 crc kubenswrapper[4696]: I0321 09:02:06.550624 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6936cdd-5be6-4162-a976-47d5df686290" path="/var/lib/kubelet/pods/f6936cdd-5be6-4162-a976-47d5df686290/volumes" Mar 21 09:02:27 crc kubenswrapper[4696]: I0321 09:02:27.450704 4696 generic.go:334] "Generic (PLEG): container finished" podID="2c75b9c9-8d94-4541-8858-0c56d8ecb400" containerID="b0d8f7d4953b05bf4a1decfa62e02d71ec519caf4b664217a87ce6de7624b690" exitCode=0 Mar 21 09:02:27 crc kubenswrapper[4696]: I0321 09:02:27.451166 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" event={"ID":"2c75b9c9-8d94-4541-8858-0c56d8ecb400","Type":"ContainerDied","Data":"b0d8f7d4953b05bf4a1decfa62e02d71ec519caf4b664217a87ce6de7624b690"} Mar 21 09:02:28 crc kubenswrapper[4696]: I0321 09:02:28.931436 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" Mar 21 09:02:28 crc kubenswrapper[4696]: I0321 09:02:28.982875 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5l5p\" (UniqueName: \"kubernetes.io/projected/2c75b9c9-8d94-4541-8858-0c56d8ecb400-kube-api-access-x5l5p\") pod \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\" (UID: \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\") " Mar 21 09:02:28 crc kubenswrapper[4696]: I0321 09:02:28.983283 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c75b9c9-8d94-4541-8858-0c56d8ecb400-ovn-combined-ca-bundle\") pod \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\" (UID: \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\") " Mar 21 09:02:28 crc kubenswrapper[4696]: I0321 09:02:28.983398 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c75b9c9-8d94-4541-8858-0c56d8ecb400-inventory\") pod \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\" (UID: \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\") " Mar 21 09:02:28 crc kubenswrapper[4696]: I0321 09:02:28.983537 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/2c75b9c9-8d94-4541-8858-0c56d8ecb400-ovncontroller-config-0\") pod \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\" (UID: \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\") " Mar 21 09:02:28 crc kubenswrapper[4696]: I0321 09:02:28.983639 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2c75b9c9-8d94-4541-8858-0c56d8ecb400-ssh-key-openstack-edpm-ipam\") pod \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\" (UID: \"2c75b9c9-8d94-4541-8858-0c56d8ecb400\") " Mar 21 09:02:28 crc kubenswrapper[4696]: I0321 09:02:28.989871 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c75b9c9-8d94-4541-8858-0c56d8ecb400-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "2c75b9c9-8d94-4541-8858-0c56d8ecb400" (UID: "2c75b9c9-8d94-4541-8858-0c56d8ecb400"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:02:28 crc kubenswrapper[4696]: I0321 09:02:28.989986 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c75b9c9-8d94-4541-8858-0c56d8ecb400-kube-api-access-x5l5p" (OuterVolumeSpecName: "kube-api-access-x5l5p") pod "2c75b9c9-8d94-4541-8858-0c56d8ecb400" (UID: "2c75b9c9-8d94-4541-8858-0c56d8ecb400"). InnerVolumeSpecName "kube-api-access-x5l5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.012605 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c75b9c9-8d94-4541-8858-0c56d8ecb400-inventory" (OuterVolumeSpecName: "inventory") pod "2c75b9c9-8d94-4541-8858-0c56d8ecb400" (UID: "2c75b9c9-8d94-4541-8858-0c56d8ecb400"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.014224 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c75b9c9-8d94-4541-8858-0c56d8ecb400-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2c75b9c9-8d94-4541-8858-0c56d8ecb400" (UID: "2c75b9c9-8d94-4541-8858-0c56d8ecb400"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.023844 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c75b9c9-8d94-4541-8858-0c56d8ecb400-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "2c75b9c9-8d94-4541-8858-0c56d8ecb400" (UID: "2c75b9c9-8d94-4541-8858-0c56d8ecb400"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.086568 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5l5p\" (UniqueName: \"kubernetes.io/projected/2c75b9c9-8d94-4541-8858-0c56d8ecb400-kube-api-access-x5l5p\") on node \"crc\" DevicePath \"\"" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.086605 4696 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c75b9c9-8d94-4541-8858-0c56d8ecb400-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.086619 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c75b9c9-8d94-4541-8858-0c56d8ecb400-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.086632 4696 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/2c75b9c9-8d94-4541-8858-0c56d8ecb400-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.086644 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2c75b9c9-8d94-4541-8858-0c56d8ecb400-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.477141 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" event={"ID":"2c75b9c9-8d94-4541-8858-0c56d8ecb400","Type":"ContainerDied","Data":"9d124e9e52a2750c16d6ab29d044bb7cac8d0e4958f30cd7fa6a182b27844272"} Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.477601 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d124e9e52a2750c16d6ab29d044bb7cac8d0e4958f30cd7fa6a182b27844272" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.477211 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-v5hcz" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.597300 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv"] Mar 21 09:02:29 crc kubenswrapper[4696]: E0321 09:02:29.597901 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c75b9c9-8d94-4541-8858-0c56d8ecb400" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.597925 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c75b9c9-8d94-4541-8858-0c56d8ecb400" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 21 09:02:29 crc kubenswrapper[4696]: E0321 09:02:29.597940 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d1e242-e372-472f-a67c-709b9d94dfe4" containerName="oc" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.597948 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d1e242-e372-472f-a67c-709b9d94dfe4" containerName="oc" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.598209 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1d1e242-e372-472f-a67c-709b9d94dfe4" containerName="oc" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.598237 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c75b9c9-8d94-4541-8858-0c56d8ecb400" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.599180 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.602157 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.602461 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.602700 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bqp9r" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.602887 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.603090 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.603303 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.615881 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv"] Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.699262 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf9gb\" (UniqueName: \"kubernetes.io/projected/76b63ef6-105a-4d99-a8bc-a8758970ffff-kube-api-access-qf9gb\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.699305 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.699399 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.699479 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.699590 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.699712 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.801454 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.801736 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf9gb\" (UniqueName: \"kubernetes.io/projected/76b63ef6-105a-4d99-a8bc-a8758970ffff-kube-api-access-qf9gb\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.801772 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.801806 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.801867 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.801950 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.807271 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.807396 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.807966 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.808902 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.814273 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.835357 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf9gb\" (UniqueName: \"kubernetes.io/projected/76b63ef6-105a-4d99-a8bc-a8758970ffff-kube-api-access-qf9gb\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:02:29 crc kubenswrapper[4696]: I0321 09:02:29.928306 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:02:30 crc kubenswrapper[4696]: I0321 09:02:30.341167 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:02:30 crc kubenswrapper[4696]: I0321 09:02:30.341318 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:02:30 crc kubenswrapper[4696]: I0321 09:02:30.457187 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv"] Mar 21 09:02:30 crc kubenswrapper[4696]: I0321 09:02:30.465444 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 09:02:30 crc kubenswrapper[4696]: I0321 09:02:30.492756 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" event={"ID":"76b63ef6-105a-4d99-a8bc-a8758970ffff","Type":"ContainerStarted","Data":"c458d9b21465f077816db563f718f06342004c35256f50d7c20d076e277d2037"} Mar 21 09:02:31 crc kubenswrapper[4696]: I0321 09:02:31.502996 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" event={"ID":"76b63ef6-105a-4d99-a8bc-a8758970ffff","Type":"ContainerStarted","Data":"b81b50c370671d59299632f810d98fcd2e99fa9a0016d288681554c5d90ab274"} Mar 21 09:02:31 crc kubenswrapper[4696]: I0321 09:02:31.520648 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" podStartSLOduration=1.980552445 podStartE2EDuration="2.520629488s" podCreationTimestamp="2026-03-21 09:02:29 +0000 UTC" firstStartedPulling="2026-03-21 09:02:30.465129597 +0000 UTC m=+2084.586010310" lastFinishedPulling="2026-03-21 09:02:31.00520664 +0000 UTC m=+2085.126087353" observedRunningTime="2026-03-21 09:02:31.517307548 +0000 UTC m=+2085.638188261" watchObservedRunningTime="2026-03-21 09:02:31.520629488 +0000 UTC m=+2085.641510201" Mar 21 09:02:49 crc kubenswrapper[4696]: I0321 09:02:49.589051 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lgqh5"] Mar 21 09:02:49 crc kubenswrapper[4696]: I0321 09:02:49.593076 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgqh5" Mar 21 09:02:49 crc kubenswrapper[4696]: I0321 09:02:49.611592 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lgqh5"] Mar 21 09:02:49 crc kubenswrapper[4696]: I0321 09:02:49.730363 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h449k\" (UniqueName: \"kubernetes.io/projected/4377db9f-6c16-40e0-a42c-8f99609be51c-kube-api-access-h449k\") pod \"redhat-operators-lgqh5\" (UID: \"4377db9f-6c16-40e0-a42c-8f99609be51c\") " pod="openshift-marketplace/redhat-operators-lgqh5" Mar 21 09:02:49 crc kubenswrapper[4696]: I0321 09:02:49.731266 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4377db9f-6c16-40e0-a42c-8f99609be51c-catalog-content\") pod \"redhat-operators-lgqh5\" (UID: \"4377db9f-6c16-40e0-a42c-8f99609be51c\") " pod="openshift-marketplace/redhat-operators-lgqh5" Mar 21 09:02:49 crc kubenswrapper[4696]: I0321 09:02:49.731753 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4377db9f-6c16-40e0-a42c-8f99609be51c-utilities\") pod \"redhat-operators-lgqh5\" (UID: \"4377db9f-6c16-40e0-a42c-8f99609be51c\") " pod="openshift-marketplace/redhat-operators-lgqh5" Mar 21 09:02:49 crc kubenswrapper[4696]: I0321 09:02:49.834299 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h449k\" (UniqueName: \"kubernetes.io/projected/4377db9f-6c16-40e0-a42c-8f99609be51c-kube-api-access-h449k\") pod \"redhat-operators-lgqh5\" (UID: \"4377db9f-6c16-40e0-a42c-8f99609be51c\") " pod="openshift-marketplace/redhat-operators-lgqh5" Mar 21 09:02:49 crc kubenswrapper[4696]: I0321 09:02:49.834442 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4377db9f-6c16-40e0-a42c-8f99609be51c-catalog-content\") pod \"redhat-operators-lgqh5\" (UID: \"4377db9f-6c16-40e0-a42c-8f99609be51c\") " pod="openshift-marketplace/redhat-operators-lgqh5" Mar 21 09:02:49 crc kubenswrapper[4696]: I0321 09:02:49.834541 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4377db9f-6c16-40e0-a42c-8f99609be51c-utilities\") pod \"redhat-operators-lgqh5\" (UID: \"4377db9f-6c16-40e0-a42c-8f99609be51c\") " pod="openshift-marketplace/redhat-operators-lgqh5" Mar 21 09:02:49 crc kubenswrapper[4696]: I0321 09:02:49.835253 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4377db9f-6c16-40e0-a42c-8f99609be51c-catalog-content\") pod \"redhat-operators-lgqh5\" (UID: \"4377db9f-6c16-40e0-a42c-8f99609be51c\") " pod="openshift-marketplace/redhat-operators-lgqh5" Mar 21 09:02:49 crc kubenswrapper[4696]: I0321 09:02:49.835381 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4377db9f-6c16-40e0-a42c-8f99609be51c-utilities\") pod \"redhat-operators-lgqh5\" (UID: \"4377db9f-6c16-40e0-a42c-8f99609be51c\") " pod="openshift-marketplace/redhat-operators-lgqh5" Mar 21 09:02:49 crc kubenswrapper[4696]: I0321 09:02:49.873908 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h449k\" (UniqueName: \"kubernetes.io/projected/4377db9f-6c16-40e0-a42c-8f99609be51c-kube-api-access-h449k\") pod \"redhat-operators-lgqh5\" (UID: \"4377db9f-6c16-40e0-a42c-8f99609be51c\") " pod="openshift-marketplace/redhat-operators-lgqh5" Mar 21 09:02:49 crc kubenswrapper[4696]: I0321 09:02:49.915223 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgqh5" Mar 21 09:02:50 crc kubenswrapper[4696]: W0321 09:02:50.408115 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4377db9f_6c16_40e0_a42c_8f99609be51c.slice/crio-3c6bde968600d42a6c89b36ea8ec8bfe432d4c16a4147e0fbde1f1a8e6af42f1 WatchSource:0}: Error finding container 3c6bde968600d42a6c89b36ea8ec8bfe432d4c16a4147e0fbde1f1a8e6af42f1: Status 404 returned error can't find the container with id 3c6bde968600d42a6c89b36ea8ec8bfe432d4c16a4147e0fbde1f1a8e6af42f1 Mar 21 09:02:50 crc kubenswrapper[4696]: I0321 09:02:50.414095 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lgqh5"] Mar 21 09:02:50 crc kubenswrapper[4696]: I0321 09:02:50.730493 4696 generic.go:334] "Generic (PLEG): container finished" podID="4377db9f-6c16-40e0-a42c-8f99609be51c" containerID="2ca15307b4f74152cde5ee4dc8e5efc669fd251fc0c8044873dfb6550a57c5ae" exitCode=0 Mar 21 09:02:50 crc kubenswrapper[4696]: I0321 09:02:50.733934 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgqh5" event={"ID":"4377db9f-6c16-40e0-a42c-8f99609be51c","Type":"ContainerDied","Data":"2ca15307b4f74152cde5ee4dc8e5efc669fd251fc0c8044873dfb6550a57c5ae"} Mar 21 09:02:50 crc kubenswrapper[4696]: I0321 09:02:50.734016 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgqh5" event={"ID":"4377db9f-6c16-40e0-a42c-8f99609be51c","Type":"ContainerStarted","Data":"3c6bde968600d42a6c89b36ea8ec8bfe432d4c16a4147e0fbde1f1a8e6af42f1"} Mar 21 09:02:51 crc kubenswrapper[4696]: I0321 09:02:51.742682 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgqh5" event={"ID":"4377db9f-6c16-40e0-a42c-8f99609be51c","Type":"ContainerStarted","Data":"6b2757086ce5edb7d068adc4001e6d97da9b2af7d6d673d56e7cc95295d22915"} Mar 21 09:02:55 crc kubenswrapper[4696]: I0321 09:02:55.777596 4696 generic.go:334] "Generic (PLEG): container finished" podID="4377db9f-6c16-40e0-a42c-8f99609be51c" containerID="6b2757086ce5edb7d068adc4001e6d97da9b2af7d6d673d56e7cc95295d22915" exitCode=0 Mar 21 09:02:55 crc kubenswrapper[4696]: I0321 09:02:55.777761 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgqh5" event={"ID":"4377db9f-6c16-40e0-a42c-8f99609be51c","Type":"ContainerDied","Data":"6b2757086ce5edb7d068adc4001e6d97da9b2af7d6d673d56e7cc95295d22915"} Mar 21 09:02:56 crc kubenswrapper[4696]: I0321 09:02:56.788887 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgqh5" event={"ID":"4377db9f-6c16-40e0-a42c-8f99609be51c","Type":"ContainerStarted","Data":"b8082c1595dffd0db729ca14c819d77c2ccf04d88816630c479da7248af59898"} Mar 21 09:02:56 crc kubenswrapper[4696]: I0321 09:02:56.811445 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lgqh5" podStartSLOduration=2.354486102 podStartE2EDuration="7.811423965s" podCreationTimestamp="2026-03-21 09:02:49 +0000 UTC" firstStartedPulling="2026-03-21 09:02:50.735067676 +0000 UTC m=+2104.855948389" lastFinishedPulling="2026-03-21 09:02:56.192005539 +0000 UTC m=+2110.312886252" observedRunningTime="2026-03-21 09:02:56.806922672 +0000 UTC m=+2110.927803385" watchObservedRunningTime="2026-03-21 09:02:56.811423965 +0000 UTC m=+2110.932304678" Mar 21 09:02:59 crc kubenswrapper[4696]: I0321 09:02:59.915906 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lgqh5" Mar 21 09:02:59 crc kubenswrapper[4696]: I0321 09:02:59.917477 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lgqh5" Mar 21 09:03:00 crc kubenswrapper[4696]: I0321 09:03:00.340945 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:03:00 crc kubenswrapper[4696]: I0321 09:03:00.341018 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:03:00 crc kubenswrapper[4696]: I0321 09:03:00.977573 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lgqh5" podUID="4377db9f-6c16-40e0-a42c-8f99609be51c" containerName="registry-server" probeResult="failure" output=< Mar 21 09:03:00 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Mar 21 09:03:00 crc kubenswrapper[4696]: > Mar 21 09:03:01 crc kubenswrapper[4696]: I0321 09:03:01.475804 4696 scope.go:117] "RemoveContainer" containerID="ad43bf346eb6430191fea0d17617da01fb6e57eca827d0006fe34f07eaa54a53" Mar 21 09:03:09 crc kubenswrapper[4696]: I0321 09:03:09.961348 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lgqh5" Mar 21 09:03:10 crc kubenswrapper[4696]: I0321 09:03:10.039581 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lgqh5" Mar 21 09:03:10 crc kubenswrapper[4696]: I0321 09:03:10.213423 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lgqh5"] Mar 21 09:03:11 crc kubenswrapper[4696]: I0321 09:03:11.960007 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lgqh5" podUID="4377db9f-6c16-40e0-a42c-8f99609be51c" containerName="registry-server" containerID="cri-o://b8082c1595dffd0db729ca14c819d77c2ccf04d88816630c479da7248af59898" gracePeriod=2 Mar 21 09:03:12 crc kubenswrapper[4696]: I0321 09:03:12.489771 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgqh5" Mar 21 09:03:12 crc kubenswrapper[4696]: I0321 09:03:12.634389 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4377db9f-6c16-40e0-a42c-8f99609be51c-catalog-content\") pod \"4377db9f-6c16-40e0-a42c-8f99609be51c\" (UID: \"4377db9f-6c16-40e0-a42c-8f99609be51c\") " Mar 21 09:03:12 crc kubenswrapper[4696]: I0321 09:03:12.634508 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h449k\" (UniqueName: \"kubernetes.io/projected/4377db9f-6c16-40e0-a42c-8f99609be51c-kube-api-access-h449k\") pod \"4377db9f-6c16-40e0-a42c-8f99609be51c\" (UID: \"4377db9f-6c16-40e0-a42c-8f99609be51c\") " Mar 21 09:03:12 crc kubenswrapper[4696]: I0321 09:03:12.634574 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4377db9f-6c16-40e0-a42c-8f99609be51c-utilities\") pod \"4377db9f-6c16-40e0-a42c-8f99609be51c\" (UID: \"4377db9f-6c16-40e0-a42c-8f99609be51c\") " Mar 21 09:03:12 crc kubenswrapper[4696]: I0321 09:03:12.635277 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4377db9f-6c16-40e0-a42c-8f99609be51c-utilities" (OuterVolumeSpecName: "utilities") pod "4377db9f-6c16-40e0-a42c-8f99609be51c" (UID: "4377db9f-6c16-40e0-a42c-8f99609be51c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:03:12 crc kubenswrapper[4696]: I0321 09:03:12.636796 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4377db9f-6c16-40e0-a42c-8f99609be51c-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 09:03:12 crc kubenswrapper[4696]: I0321 09:03:12.646076 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4377db9f-6c16-40e0-a42c-8f99609be51c-kube-api-access-h449k" (OuterVolumeSpecName: "kube-api-access-h449k") pod "4377db9f-6c16-40e0-a42c-8f99609be51c" (UID: "4377db9f-6c16-40e0-a42c-8f99609be51c"). InnerVolumeSpecName "kube-api-access-h449k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:03:12 crc kubenswrapper[4696]: I0321 09:03:12.738708 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h449k\" (UniqueName: \"kubernetes.io/projected/4377db9f-6c16-40e0-a42c-8f99609be51c-kube-api-access-h449k\") on node \"crc\" DevicePath \"\"" Mar 21 09:03:12 crc kubenswrapper[4696]: I0321 09:03:12.801431 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4377db9f-6c16-40e0-a42c-8f99609be51c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4377db9f-6c16-40e0-a42c-8f99609be51c" (UID: "4377db9f-6c16-40e0-a42c-8f99609be51c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:03:12 crc kubenswrapper[4696]: I0321 09:03:12.841036 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4377db9f-6c16-40e0-a42c-8f99609be51c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 09:03:12 crc kubenswrapper[4696]: I0321 09:03:12.973566 4696 generic.go:334] "Generic (PLEG): container finished" podID="4377db9f-6c16-40e0-a42c-8f99609be51c" containerID="b8082c1595dffd0db729ca14c819d77c2ccf04d88816630c479da7248af59898" exitCode=0 Mar 21 09:03:12 crc kubenswrapper[4696]: I0321 09:03:12.973862 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgqh5" event={"ID":"4377db9f-6c16-40e0-a42c-8f99609be51c","Type":"ContainerDied","Data":"b8082c1595dffd0db729ca14c819d77c2ccf04d88816630c479da7248af59898"} Mar 21 09:03:12 crc kubenswrapper[4696]: I0321 09:03:12.973893 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgqh5" event={"ID":"4377db9f-6c16-40e0-a42c-8f99609be51c","Type":"ContainerDied","Data":"3c6bde968600d42a6c89b36ea8ec8bfe432d4c16a4147e0fbde1f1a8e6af42f1"} Mar 21 09:03:12 crc kubenswrapper[4696]: I0321 09:03:12.973912 4696 scope.go:117] "RemoveContainer" containerID="b8082c1595dffd0db729ca14c819d77c2ccf04d88816630c479da7248af59898" Mar 21 09:03:12 crc kubenswrapper[4696]: I0321 09:03:12.974030 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgqh5" Mar 21 09:03:13 crc kubenswrapper[4696]: I0321 09:03:13.008741 4696 scope.go:117] "RemoveContainer" containerID="6b2757086ce5edb7d068adc4001e6d97da9b2af7d6d673d56e7cc95295d22915" Mar 21 09:03:13 crc kubenswrapper[4696]: I0321 09:03:13.016278 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lgqh5"] Mar 21 09:03:13 crc kubenswrapper[4696]: I0321 09:03:13.025902 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lgqh5"] Mar 21 09:03:13 crc kubenswrapper[4696]: I0321 09:03:13.032378 4696 scope.go:117] "RemoveContainer" containerID="2ca15307b4f74152cde5ee4dc8e5efc669fd251fc0c8044873dfb6550a57c5ae" Mar 21 09:03:13 crc kubenswrapper[4696]: E0321 09:03:13.064765 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4377db9f_6c16_40e0_a42c_8f99609be51c.slice\": RecentStats: unable to find data in memory cache]" Mar 21 09:03:13 crc kubenswrapper[4696]: I0321 09:03:13.090409 4696 scope.go:117] "RemoveContainer" containerID="b8082c1595dffd0db729ca14c819d77c2ccf04d88816630c479da7248af59898" Mar 21 09:03:13 crc kubenswrapper[4696]: E0321 09:03:13.091797 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8082c1595dffd0db729ca14c819d77c2ccf04d88816630c479da7248af59898\": container with ID starting with b8082c1595dffd0db729ca14c819d77c2ccf04d88816630c479da7248af59898 not found: ID does not exist" containerID="b8082c1595dffd0db729ca14c819d77c2ccf04d88816630c479da7248af59898" Mar 21 09:03:13 crc kubenswrapper[4696]: I0321 09:03:13.091837 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8082c1595dffd0db729ca14c819d77c2ccf04d88816630c479da7248af59898"} err="failed to get container status \"b8082c1595dffd0db729ca14c819d77c2ccf04d88816630c479da7248af59898\": rpc error: code = NotFound desc = could not find container \"b8082c1595dffd0db729ca14c819d77c2ccf04d88816630c479da7248af59898\": container with ID starting with b8082c1595dffd0db729ca14c819d77c2ccf04d88816630c479da7248af59898 not found: ID does not exist" Mar 21 09:03:13 crc kubenswrapper[4696]: I0321 09:03:13.091857 4696 scope.go:117] "RemoveContainer" containerID="6b2757086ce5edb7d068adc4001e6d97da9b2af7d6d673d56e7cc95295d22915" Mar 21 09:03:13 crc kubenswrapper[4696]: E0321 09:03:13.092114 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b2757086ce5edb7d068adc4001e6d97da9b2af7d6d673d56e7cc95295d22915\": container with ID starting with 6b2757086ce5edb7d068adc4001e6d97da9b2af7d6d673d56e7cc95295d22915 not found: ID does not exist" containerID="6b2757086ce5edb7d068adc4001e6d97da9b2af7d6d673d56e7cc95295d22915" Mar 21 09:03:13 crc kubenswrapper[4696]: I0321 09:03:13.092136 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b2757086ce5edb7d068adc4001e6d97da9b2af7d6d673d56e7cc95295d22915"} err="failed to get container status \"6b2757086ce5edb7d068adc4001e6d97da9b2af7d6d673d56e7cc95295d22915\": rpc error: code = NotFound desc = could not find container \"6b2757086ce5edb7d068adc4001e6d97da9b2af7d6d673d56e7cc95295d22915\": container with ID starting with 6b2757086ce5edb7d068adc4001e6d97da9b2af7d6d673d56e7cc95295d22915 not found: ID does not exist" Mar 21 09:03:13 crc kubenswrapper[4696]: I0321 09:03:13.092150 4696 scope.go:117] "RemoveContainer" containerID="2ca15307b4f74152cde5ee4dc8e5efc669fd251fc0c8044873dfb6550a57c5ae" Mar 21 09:03:13 crc kubenswrapper[4696]: E0321 09:03:13.092539 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ca15307b4f74152cde5ee4dc8e5efc669fd251fc0c8044873dfb6550a57c5ae\": container with ID starting with 2ca15307b4f74152cde5ee4dc8e5efc669fd251fc0c8044873dfb6550a57c5ae not found: ID does not exist" containerID="2ca15307b4f74152cde5ee4dc8e5efc669fd251fc0c8044873dfb6550a57c5ae" Mar 21 09:03:13 crc kubenswrapper[4696]: I0321 09:03:13.092562 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ca15307b4f74152cde5ee4dc8e5efc669fd251fc0c8044873dfb6550a57c5ae"} err="failed to get container status \"2ca15307b4f74152cde5ee4dc8e5efc669fd251fc0c8044873dfb6550a57c5ae\": rpc error: code = NotFound desc = could not find container \"2ca15307b4f74152cde5ee4dc8e5efc669fd251fc0c8044873dfb6550a57c5ae\": container with ID starting with 2ca15307b4f74152cde5ee4dc8e5efc669fd251fc0c8044873dfb6550a57c5ae not found: ID does not exist" Mar 21 09:03:14 crc kubenswrapper[4696]: I0321 09:03:14.547564 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4377db9f-6c16-40e0-a42c-8f99609be51c" path="/var/lib/kubelet/pods/4377db9f-6c16-40e0-a42c-8f99609be51c/volumes" Mar 21 09:03:19 crc kubenswrapper[4696]: I0321 09:03:19.034101 4696 generic.go:334] "Generic (PLEG): container finished" podID="76b63ef6-105a-4d99-a8bc-a8758970ffff" containerID="b81b50c370671d59299632f810d98fcd2e99fa9a0016d288681554c5d90ab274" exitCode=0 Mar 21 09:03:19 crc kubenswrapper[4696]: I0321 09:03:19.034196 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" event={"ID":"76b63ef6-105a-4d99-a8bc-a8758970ffff","Type":"ContainerDied","Data":"b81b50c370671d59299632f810d98fcd2e99fa9a0016d288681554c5d90ab274"} Mar 21 09:03:21 crc kubenswrapper[4696]: I0321 09:03:21.057142 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:03:21 crc kubenswrapper[4696]: I0321 09:03:21.057138 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" event={"ID":"76b63ef6-105a-4d99-a8bc-a8758970ffff","Type":"ContainerDied","Data":"c458d9b21465f077816db563f718f06342004c35256f50d7c20d076e277d2037"} Mar 21 09:03:21 crc kubenswrapper[4696]: I0321 09:03:21.057925 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c458d9b21465f077816db563f718f06342004c35256f50d7c20d076e277d2037" Mar 21 09:03:21 crc kubenswrapper[4696]: I0321 09:03:21.131121 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-nova-metadata-neutron-config-0\") pod \"76b63ef6-105a-4d99-a8bc-a8758970ffff\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " Mar 21 09:03:21 crc kubenswrapper[4696]: I0321 09:03:21.131195 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qf9gb\" (UniqueName: \"kubernetes.io/projected/76b63ef6-105a-4d99-a8bc-a8758970ffff-kube-api-access-qf9gb\") pod \"76b63ef6-105a-4d99-a8bc-a8758970ffff\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " Mar 21 09:03:21 crc kubenswrapper[4696]: I0321 09:03:21.131222 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-neutron-metadata-combined-ca-bundle\") pod \"76b63ef6-105a-4d99-a8bc-a8758970ffff\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " Mar 21 09:03:21 crc kubenswrapper[4696]: I0321 09:03:21.131288 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-ssh-key-openstack-edpm-ipam\") pod \"76b63ef6-105a-4d99-a8bc-a8758970ffff\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " Mar 21 09:03:21 crc kubenswrapper[4696]: I0321 09:03:21.131316 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-neutron-ovn-metadata-agent-neutron-config-0\") pod \"76b63ef6-105a-4d99-a8bc-a8758970ffff\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " Mar 21 09:03:21 crc kubenswrapper[4696]: I0321 09:03:21.131381 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-inventory\") pod \"76b63ef6-105a-4d99-a8bc-a8758970ffff\" (UID: \"76b63ef6-105a-4d99-a8bc-a8758970ffff\") " Mar 21 09:03:21 crc kubenswrapper[4696]: I0321 09:03:21.137767 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "76b63ef6-105a-4d99-a8bc-a8758970ffff" (UID: "76b63ef6-105a-4d99-a8bc-a8758970ffff"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:03:21 crc kubenswrapper[4696]: I0321 09:03:21.137781 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76b63ef6-105a-4d99-a8bc-a8758970ffff-kube-api-access-qf9gb" (OuterVolumeSpecName: "kube-api-access-qf9gb") pod "76b63ef6-105a-4d99-a8bc-a8758970ffff" (UID: "76b63ef6-105a-4d99-a8bc-a8758970ffff"). InnerVolumeSpecName "kube-api-access-qf9gb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:03:21 crc kubenswrapper[4696]: I0321 09:03:21.166867 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "76b63ef6-105a-4d99-a8bc-a8758970ffff" (UID: "76b63ef6-105a-4d99-a8bc-a8758970ffff"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:03:21 crc kubenswrapper[4696]: I0321 09:03:21.167228 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "76b63ef6-105a-4d99-a8bc-a8758970ffff" (UID: "76b63ef6-105a-4d99-a8bc-a8758970ffff"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:03:21 crc kubenswrapper[4696]: I0321 09:03:21.176774 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-inventory" (OuterVolumeSpecName: "inventory") pod "76b63ef6-105a-4d99-a8bc-a8758970ffff" (UID: "76b63ef6-105a-4d99-a8bc-a8758970ffff"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:03:21 crc kubenswrapper[4696]: I0321 09:03:21.176839 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "76b63ef6-105a-4d99-a8bc-a8758970ffff" (UID: "76b63ef6-105a-4d99-a8bc-a8758970ffff"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:03:21 crc kubenswrapper[4696]: I0321 09:03:21.233310 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qf9gb\" (UniqueName: \"kubernetes.io/projected/76b63ef6-105a-4d99-a8bc-a8758970ffff-kube-api-access-qf9gb\") on node \"crc\" DevicePath \"\"" Mar 21 09:03:21 crc kubenswrapper[4696]: I0321 09:03:21.233350 4696 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 09:03:21 crc kubenswrapper[4696]: I0321 09:03:21.233363 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 09:03:21 crc kubenswrapper[4696]: I0321 09:03:21.233375 4696 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 21 09:03:21 crc kubenswrapper[4696]: I0321 09:03:21.233387 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 09:03:21 crc kubenswrapper[4696]: I0321 09:03:21.233397 4696 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/76b63ef6-105a-4d99-a8bc-a8758970ffff-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.063915 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.232203 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d"] Mar 21 09:03:22 crc kubenswrapper[4696]: E0321 09:03:22.232867 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4377db9f-6c16-40e0-a42c-8f99609be51c" containerName="extract-content" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.232887 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4377db9f-6c16-40e0-a42c-8f99609be51c" containerName="extract-content" Mar 21 09:03:22 crc kubenswrapper[4696]: E0321 09:03:22.232915 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4377db9f-6c16-40e0-a42c-8f99609be51c" containerName="registry-server" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.232921 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4377db9f-6c16-40e0-a42c-8f99609be51c" containerName="registry-server" Mar 21 09:03:22 crc kubenswrapper[4696]: E0321 09:03:22.232939 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4377db9f-6c16-40e0-a42c-8f99609be51c" containerName="extract-utilities" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.232945 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4377db9f-6c16-40e0-a42c-8f99609be51c" containerName="extract-utilities" Mar 21 09:03:22 crc kubenswrapper[4696]: E0321 09:03:22.232954 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76b63ef6-105a-4d99-a8bc-a8758970ffff" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.232961 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="76b63ef6-105a-4d99-a8bc-a8758970ffff" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.233132 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="4377db9f-6c16-40e0-a42c-8f99609be51c" containerName="registry-server" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.233146 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="76b63ef6-105a-4d99-a8bc-a8758970ffff" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.233878 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.236993 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.237207 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bqp9r" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.237256 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.237328 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.238133 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.244795 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d"] Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.354969 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d\" (UID: \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.355065 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d\" (UID: \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.355111 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrj5v\" (UniqueName: \"kubernetes.io/projected/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-kube-api-access-rrj5v\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d\" (UID: \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.355146 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d\" (UID: \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.355270 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d\" (UID: \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.457214 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d\" (UID: \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.457347 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrj5v\" (UniqueName: \"kubernetes.io/projected/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-kube-api-access-rrj5v\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d\" (UID: \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.457400 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d\" (UID: \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.457450 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d\" (UID: \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.457556 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d\" (UID: \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.464131 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d\" (UID: \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.465039 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d\" (UID: \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.465736 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d\" (UID: \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.469772 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d\" (UID: \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.472947 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrj5v\" (UniqueName: \"kubernetes.io/projected/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-kube-api-access-rrj5v\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d\" (UID: \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" Mar 21 09:03:22 crc kubenswrapper[4696]: I0321 09:03:22.549250 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" Mar 21 09:03:23 crc kubenswrapper[4696]: I0321 09:03:23.177399 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d"] Mar 21 09:03:24 crc kubenswrapper[4696]: I0321 09:03:24.087238 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" event={"ID":"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde","Type":"ContainerStarted","Data":"0e305f921e2b2b2aa6bba2e1428e4e067e0b0b8ca3ff53575edabe788eee8b96"} Mar 21 09:03:25 crc kubenswrapper[4696]: I0321 09:03:25.106721 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" event={"ID":"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde","Type":"ContainerStarted","Data":"fb47bfe25466eaaa65a67be0c30bc309e37f7c418327b2d31077b82071a5b6d0"} Mar 21 09:03:25 crc kubenswrapper[4696]: I0321 09:03:25.136409 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" podStartSLOduration=2.396108781 podStartE2EDuration="3.136381512s" podCreationTimestamp="2026-03-21 09:03:22 +0000 UTC" firstStartedPulling="2026-03-21 09:03:23.196784203 +0000 UTC m=+2137.317664916" lastFinishedPulling="2026-03-21 09:03:23.937056934 +0000 UTC m=+2138.057937647" observedRunningTime="2026-03-21 09:03:25.135764295 +0000 UTC m=+2139.256645008" watchObservedRunningTime="2026-03-21 09:03:25.136381512 +0000 UTC m=+2139.257262225" Mar 21 09:03:30 crc kubenswrapper[4696]: I0321 09:03:30.341291 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:03:30 crc kubenswrapper[4696]: I0321 09:03:30.343379 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:03:30 crc kubenswrapper[4696]: I0321 09:03:30.343561 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 09:03:30 crc kubenswrapper[4696]: I0321 09:03:30.344871 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3187c3e490f50d1673fbeb2f54747e27d740a56d231e0d24142915c7db238efa"} pod="openshift-machine-config-operator/machine-config-daemon-z7srw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 09:03:30 crc kubenswrapper[4696]: I0321 09:03:30.345185 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" containerID="cri-o://3187c3e490f50d1673fbeb2f54747e27d740a56d231e0d24142915c7db238efa" gracePeriod=600 Mar 21 09:03:31 crc kubenswrapper[4696]: I0321 09:03:31.167572 4696 generic.go:334] "Generic (PLEG): container finished" podID="daaf227a-2305-495e-8495-a280abcd8e10" containerID="3187c3e490f50d1673fbeb2f54747e27d740a56d231e0d24142915c7db238efa" exitCode=0 Mar 21 09:03:31 crc kubenswrapper[4696]: I0321 09:03:31.167672 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerDied","Data":"3187c3e490f50d1673fbeb2f54747e27d740a56d231e0d24142915c7db238efa"} Mar 21 09:03:31 crc kubenswrapper[4696]: I0321 09:03:31.168136 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerStarted","Data":"d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263"} Mar 21 09:03:31 crc kubenswrapper[4696]: I0321 09:03:31.168159 4696 scope.go:117] "RemoveContainer" containerID="a80f38ecb84daef7900ac0878baf981e2fc1b724c0da04bde7e8d993712e4959" Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.145153 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568064-lq59p"] Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.147551 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568064-lq59p" Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.153274 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.153385 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.153630 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.156867 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568064-lq59p"] Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.261541 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hnrv\" (UniqueName: \"kubernetes.io/projected/d52bfab3-d526-433b-86ac-3213f08251d6-kube-api-access-7hnrv\") pod \"auto-csr-approver-29568064-lq59p\" (UID: \"d52bfab3-d526-433b-86ac-3213f08251d6\") " pod="openshift-infra/auto-csr-approver-29568064-lq59p" Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.363645 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hnrv\" (UniqueName: \"kubernetes.io/projected/d52bfab3-d526-433b-86ac-3213f08251d6-kube-api-access-7hnrv\") pod \"auto-csr-approver-29568064-lq59p\" (UID: \"d52bfab3-d526-433b-86ac-3213f08251d6\") " pod="openshift-infra/auto-csr-approver-29568064-lq59p" Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.389929 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hnrv\" (UniqueName: \"kubernetes.io/projected/d52bfab3-d526-433b-86ac-3213f08251d6-kube-api-access-7hnrv\") pod \"auto-csr-approver-29568064-lq59p\" (UID: \"d52bfab3-d526-433b-86ac-3213f08251d6\") " pod="openshift-infra/auto-csr-approver-29568064-lq59p" Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.448704 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lxndl"] Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.452619 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lxndl" Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.471085 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568064-lq59p" Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.474706 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lxndl"] Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.567850 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-789qg\" (UniqueName: \"kubernetes.io/projected/7fc4435a-7193-4c62-99c4-008ac4555225-kube-api-access-789qg\") pod \"certified-operators-lxndl\" (UID: \"7fc4435a-7193-4c62-99c4-008ac4555225\") " pod="openshift-marketplace/certified-operators-lxndl" Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.567962 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc4435a-7193-4c62-99c4-008ac4555225-catalog-content\") pod \"certified-operators-lxndl\" (UID: \"7fc4435a-7193-4c62-99c4-008ac4555225\") " pod="openshift-marketplace/certified-operators-lxndl" Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.568468 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc4435a-7193-4c62-99c4-008ac4555225-utilities\") pod \"certified-operators-lxndl\" (UID: \"7fc4435a-7193-4c62-99c4-008ac4555225\") " pod="openshift-marketplace/certified-operators-lxndl" Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.671244 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc4435a-7193-4c62-99c4-008ac4555225-utilities\") pod \"certified-operators-lxndl\" (UID: \"7fc4435a-7193-4c62-99c4-008ac4555225\") " pod="openshift-marketplace/certified-operators-lxndl" Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.671546 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-789qg\" (UniqueName: \"kubernetes.io/projected/7fc4435a-7193-4c62-99c4-008ac4555225-kube-api-access-789qg\") pod \"certified-operators-lxndl\" (UID: \"7fc4435a-7193-4c62-99c4-008ac4555225\") " pod="openshift-marketplace/certified-operators-lxndl" Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.671603 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc4435a-7193-4c62-99c4-008ac4555225-catalog-content\") pod \"certified-operators-lxndl\" (UID: \"7fc4435a-7193-4c62-99c4-008ac4555225\") " pod="openshift-marketplace/certified-operators-lxndl" Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.673525 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc4435a-7193-4c62-99c4-008ac4555225-utilities\") pod \"certified-operators-lxndl\" (UID: \"7fc4435a-7193-4c62-99c4-008ac4555225\") " pod="openshift-marketplace/certified-operators-lxndl" Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.675385 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc4435a-7193-4c62-99c4-008ac4555225-catalog-content\") pod \"certified-operators-lxndl\" (UID: \"7fc4435a-7193-4c62-99c4-008ac4555225\") " pod="openshift-marketplace/certified-operators-lxndl" Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.693215 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-789qg\" (UniqueName: \"kubernetes.io/projected/7fc4435a-7193-4c62-99c4-008ac4555225-kube-api-access-789qg\") pod \"certified-operators-lxndl\" (UID: \"7fc4435a-7193-4c62-99c4-008ac4555225\") " pod="openshift-marketplace/certified-operators-lxndl" Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.785483 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lxndl" Mar 21 09:04:00 crc kubenswrapper[4696]: I0321 09:04:00.993707 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568064-lq59p"] Mar 21 09:04:01 crc kubenswrapper[4696]: I0321 09:04:01.398685 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lxndl"] Mar 21 09:04:01 crc kubenswrapper[4696]: I0321 09:04:01.459066 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568064-lq59p" event={"ID":"d52bfab3-d526-433b-86ac-3213f08251d6","Type":"ContainerStarted","Data":"544dea521db245f4946a06ef428a575d450a86039de7c3a633eebf1bc851c2ff"} Mar 21 09:04:01 crc kubenswrapper[4696]: I0321 09:04:01.461082 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxndl" event={"ID":"7fc4435a-7193-4c62-99c4-008ac4555225","Type":"ContainerStarted","Data":"5666f9b5432845ede94fa66e31b8de6d97ec211cc65a33886597a292db8104f0"} Mar 21 09:04:02 crc kubenswrapper[4696]: I0321 09:04:02.472119 4696 generic.go:334] "Generic (PLEG): container finished" podID="7fc4435a-7193-4c62-99c4-008ac4555225" containerID="ffb65f0f246951fcac164ce9a2886d85e3d193155e6a753292814e4af9f7657a" exitCode=0 Mar 21 09:04:02 crc kubenswrapper[4696]: I0321 09:04:02.472199 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxndl" event={"ID":"7fc4435a-7193-4c62-99c4-008ac4555225","Type":"ContainerDied","Data":"ffb65f0f246951fcac164ce9a2886d85e3d193155e6a753292814e4af9f7657a"} Mar 21 09:04:02 crc kubenswrapper[4696]: I0321 09:04:02.474807 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568064-lq59p" event={"ID":"d52bfab3-d526-433b-86ac-3213f08251d6","Type":"ContainerStarted","Data":"bf3c73e483decd43c6b32dfde33125cf9a48de112affaa939958925a8bb1c3c3"} Mar 21 09:04:02 crc kubenswrapper[4696]: I0321 09:04:02.518657 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29568064-lq59p" podStartSLOduration=1.4246062959999999 podStartE2EDuration="2.518631074s" podCreationTimestamp="2026-03-21 09:04:00 +0000 UTC" firstStartedPulling="2026-03-21 09:04:01.000521734 +0000 UTC m=+2175.121402457" lastFinishedPulling="2026-03-21 09:04:02.094546522 +0000 UTC m=+2176.215427235" observedRunningTime="2026-03-21 09:04:02.508566416 +0000 UTC m=+2176.629447129" watchObservedRunningTime="2026-03-21 09:04:02.518631074 +0000 UTC m=+2176.639511797" Mar 21 09:04:03 crc kubenswrapper[4696]: I0321 09:04:03.484997 4696 generic.go:334] "Generic (PLEG): container finished" podID="d52bfab3-d526-433b-86ac-3213f08251d6" containerID="bf3c73e483decd43c6b32dfde33125cf9a48de112affaa939958925a8bb1c3c3" exitCode=0 Mar 21 09:04:03 crc kubenswrapper[4696]: I0321 09:04:03.485139 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568064-lq59p" event={"ID":"d52bfab3-d526-433b-86ac-3213f08251d6","Type":"ContainerDied","Data":"bf3c73e483decd43c6b32dfde33125cf9a48de112affaa939958925a8bb1c3c3"} Mar 21 09:04:03 crc kubenswrapper[4696]: I0321 09:04:03.487373 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxndl" event={"ID":"7fc4435a-7193-4c62-99c4-008ac4555225","Type":"ContainerStarted","Data":"f1983d8f1678cf6dae3be41458d2ad5ef5379765741d807a5403a1ac526a8c1e"} Mar 21 09:04:04 crc kubenswrapper[4696]: E0321 09:04:04.329924 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fc4435a_7193_4c62_99c4_008ac4555225.slice/crio-conmon-f1983d8f1678cf6dae3be41458d2ad5ef5379765741d807a5403a1ac526a8c1e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fc4435a_7193_4c62_99c4_008ac4555225.slice/crio-f1983d8f1678cf6dae3be41458d2ad5ef5379765741d807a5403a1ac526a8c1e.scope\": RecentStats: unable to find data in memory cache]" Mar 21 09:04:04 crc kubenswrapper[4696]: I0321 09:04:04.497655 4696 generic.go:334] "Generic (PLEG): container finished" podID="7fc4435a-7193-4c62-99c4-008ac4555225" containerID="f1983d8f1678cf6dae3be41458d2ad5ef5379765741d807a5403a1ac526a8c1e" exitCode=0 Mar 21 09:04:04 crc kubenswrapper[4696]: I0321 09:04:04.497755 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxndl" event={"ID":"7fc4435a-7193-4c62-99c4-008ac4555225","Type":"ContainerDied","Data":"f1983d8f1678cf6dae3be41458d2ad5ef5379765741d807a5403a1ac526a8c1e"} Mar 21 09:04:04 crc kubenswrapper[4696]: I0321 09:04:04.920729 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568064-lq59p" Mar 21 09:04:04 crc kubenswrapper[4696]: I0321 09:04:04.980012 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hnrv\" (UniqueName: \"kubernetes.io/projected/d52bfab3-d526-433b-86ac-3213f08251d6-kube-api-access-7hnrv\") pod \"d52bfab3-d526-433b-86ac-3213f08251d6\" (UID: \"d52bfab3-d526-433b-86ac-3213f08251d6\") " Mar 21 09:04:04 crc kubenswrapper[4696]: I0321 09:04:04.986650 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d52bfab3-d526-433b-86ac-3213f08251d6-kube-api-access-7hnrv" (OuterVolumeSpecName: "kube-api-access-7hnrv") pod "d52bfab3-d526-433b-86ac-3213f08251d6" (UID: "d52bfab3-d526-433b-86ac-3213f08251d6"). InnerVolumeSpecName "kube-api-access-7hnrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:04:05 crc kubenswrapper[4696]: I0321 09:04:05.083060 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hnrv\" (UniqueName: \"kubernetes.io/projected/d52bfab3-d526-433b-86ac-3213f08251d6-kube-api-access-7hnrv\") on node \"crc\" DevicePath \"\"" Mar 21 09:04:05 crc kubenswrapper[4696]: I0321 09:04:05.509175 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568064-lq59p" Mar 21 09:04:05 crc kubenswrapper[4696]: I0321 09:04:05.509160 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568064-lq59p" event={"ID":"d52bfab3-d526-433b-86ac-3213f08251d6","Type":"ContainerDied","Data":"544dea521db245f4946a06ef428a575d450a86039de7c3a633eebf1bc851c2ff"} Mar 21 09:04:05 crc kubenswrapper[4696]: I0321 09:04:05.509594 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="544dea521db245f4946a06ef428a575d450a86039de7c3a633eebf1bc851c2ff" Mar 21 09:04:05 crc kubenswrapper[4696]: I0321 09:04:05.512391 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxndl" event={"ID":"7fc4435a-7193-4c62-99c4-008ac4555225","Type":"ContainerStarted","Data":"e31e662f10aa5d9bb58b12d779c45f47f7ed9a41e6daffa138948bf95e103feb"} Mar 21 09:04:05 crc kubenswrapper[4696]: I0321 09:04:05.534846 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lxndl" podStartSLOduration=2.9852358900000002 podStartE2EDuration="5.534790188s" podCreationTimestamp="2026-03-21 09:04:00 +0000 UTC" firstStartedPulling="2026-03-21 09:04:02.473864196 +0000 UTC m=+2176.594744909" lastFinishedPulling="2026-03-21 09:04:05.023418494 +0000 UTC m=+2179.144299207" observedRunningTime="2026-03-21 09:04:05.528685909 +0000 UTC m=+2179.649566642" watchObservedRunningTime="2026-03-21 09:04:05.534790188 +0000 UTC m=+2179.655670901" Mar 21 09:04:06 crc kubenswrapper[4696]: I0321 09:04:06.037115 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568058-9z5qs"] Mar 21 09:04:06 crc kubenswrapper[4696]: I0321 09:04:06.052709 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568058-9z5qs"] Mar 21 09:04:06 crc kubenswrapper[4696]: I0321 09:04:06.547216 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6da51675-21b7-4468-b86f-28e19ec86020" path="/var/lib/kubelet/pods/6da51675-21b7-4468-b86f-28e19ec86020/volumes" Mar 21 09:04:10 crc kubenswrapper[4696]: I0321 09:04:10.786349 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lxndl" Mar 21 09:04:10 crc kubenswrapper[4696]: I0321 09:04:10.786974 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lxndl" Mar 21 09:04:10 crc kubenswrapper[4696]: I0321 09:04:10.839629 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lxndl" Mar 21 09:04:11 crc kubenswrapper[4696]: I0321 09:04:11.634464 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lxndl" Mar 21 09:04:11 crc kubenswrapper[4696]: I0321 09:04:11.695025 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lxndl"] Mar 21 09:04:13 crc kubenswrapper[4696]: I0321 09:04:13.593103 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lxndl" podUID="7fc4435a-7193-4c62-99c4-008ac4555225" containerName="registry-server" containerID="cri-o://e31e662f10aa5d9bb58b12d779c45f47f7ed9a41e6daffa138948bf95e103feb" gracePeriod=2 Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.093261 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lxndl" Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.187515 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc4435a-7193-4c62-99c4-008ac4555225-catalog-content\") pod \"7fc4435a-7193-4c62-99c4-008ac4555225\" (UID: \"7fc4435a-7193-4c62-99c4-008ac4555225\") " Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.187596 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-789qg\" (UniqueName: \"kubernetes.io/projected/7fc4435a-7193-4c62-99c4-008ac4555225-kube-api-access-789qg\") pod \"7fc4435a-7193-4c62-99c4-008ac4555225\" (UID: \"7fc4435a-7193-4c62-99c4-008ac4555225\") " Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.187692 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc4435a-7193-4c62-99c4-008ac4555225-utilities\") pod \"7fc4435a-7193-4c62-99c4-008ac4555225\" (UID: \"7fc4435a-7193-4c62-99c4-008ac4555225\") " Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.188591 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fc4435a-7193-4c62-99c4-008ac4555225-utilities" (OuterVolumeSpecName: "utilities") pod "7fc4435a-7193-4c62-99c4-008ac4555225" (UID: "7fc4435a-7193-4c62-99c4-008ac4555225"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.193293 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fc4435a-7193-4c62-99c4-008ac4555225-kube-api-access-789qg" (OuterVolumeSpecName: "kube-api-access-789qg") pod "7fc4435a-7193-4c62-99c4-008ac4555225" (UID: "7fc4435a-7193-4c62-99c4-008ac4555225"). InnerVolumeSpecName "kube-api-access-789qg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.237173 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fc4435a-7193-4c62-99c4-008ac4555225-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7fc4435a-7193-4c62-99c4-008ac4555225" (UID: "7fc4435a-7193-4c62-99c4-008ac4555225"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.289906 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc4435a-7193-4c62-99c4-008ac4555225-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.289939 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-789qg\" (UniqueName: \"kubernetes.io/projected/7fc4435a-7193-4c62-99c4-008ac4555225-kube-api-access-789qg\") on node \"crc\" DevicePath \"\"" Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.289951 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc4435a-7193-4c62-99c4-008ac4555225-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 09:04:14 crc kubenswrapper[4696]: E0321 09:04:14.587477 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fc4435a_7193_4c62_99c4_008ac4555225.slice/crio-5666f9b5432845ede94fa66e31b8de6d97ec211cc65a33886597a292db8104f0\": RecentStats: unable to find data in memory cache]" Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.611051 4696 generic.go:334] "Generic (PLEG): container finished" podID="7fc4435a-7193-4c62-99c4-008ac4555225" containerID="e31e662f10aa5d9bb58b12d779c45f47f7ed9a41e6daffa138948bf95e103feb" exitCode=0 Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.611094 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxndl" event={"ID":"7fc4435a-7193-4c62-99c4-008ac4555225","Type":"ContainerDied","Data":"e31e662f10aa5d9bb58b12d779c45f47f7ed9a41e6daffa138948bf95e103feb"} Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.611125 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxndl" event={"ID":"7fc4435a-7193-4c62-99c4-008ac4555225","Type":"ContainerDied","Data":"5666f9b5432845ede94fa66e31b8de6d97ec211cc65a33886597a292db8104f0"} Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.611135 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lxndl" Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.611142 4696 scope.go:117] "RemoveContainer" containerID="e31e662f10aa5d9bb58b12d779c45f47f7ed9a41e6daffa138948bf95e103feb" Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.635597 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lxndl"] Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.645759 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lxndl"] Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.652290 4696 scope.go:117] "RemoveContainer" containerID="f1983d8f1678cf6dae3be41458d2ad5ef5379765741d807a5403a1ac526a8c1e" Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.676432 4696 scope.go:117] "RemoveContainer" containerID="ffb65f0f246951fcac164ce9a2886d85e3d193155e6a753292814e4af9f7657a" Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.739739 4696 scope.go:117] "RemoveContainer" containerID="e31e662f10aa5d9bb58b12d779c45f47f7ed9a41e6daffa138948bf95e103feb" Mar 21 09:04:14 crc kubenswrapper[4696]: E0321 09:04:14.740120 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e31e662f10aa5d9bb58b12d779c45f47f7ed9a41e6daffa138948bf95e103feb\": container with ID starting with e31e662f10aa5d9bb58b12d779c45f47f7ed9a41e6daffa138948bf95e103feb not found: ID does not exist" containerID="e31e662f10aa5d9bb58b12d779c45f47f7ed9a41e6daffa138948bf95e103feb" Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.740156 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e31e662f10aa5d9bb58b12d779c45f47f7ed9a41e6daffa138948bf95e103feb"} err="failed to get container status \"e31e662f10aa5d9bb58b12d779c45f47f7ed9a41e6daffa138948bf95e103feb\": rpc error: code = NotFound desc = could not find container \"e31e662f10aa5d9bb58b12d779c45f47f7ed9a41e6daffa138948bf95e103feb\": container with ID starting with e31e662f10aa5d9bb58b12d779c45f47f7ed9a41e6daffa138948bf95e103feb not found: ID does not exist" Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.740177 4696 scope.go:117] "RemoveContainer" containerID="f1983d8f1678cf6dae3be41458d2ad5ef5379765741d807a5403a1ac526a8c1e" Mar 21 09:04:14 crc kubenswrapper[4696]: E0321 09:04:14.740572 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1983d8f1678cf6dae3be41458d2ad5ef5379765741d807a5403a1ac526a8c1e\": container with ID starting with f1983d8f1678cf6dae3be41458d2ad5ef5379765741d807a5403a1ac526a8c1e not found: ID does not exist" containerID="f1983d8f1678cf6dae3be41458d2ad5ef5379765741d807a5403a1ac526a8c1e" Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.740613 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1983d8f1678cf6dae3be41458d2ad5ef5379765741d807a5403a1ac526a8c1e"} err="failed to get container status \"f1983d8f1678cf6dae3be41458d2ad5ef5379765741d807a5403a1ac526a8c1e\": rpc error: code = NotFound desc = could not find container \"f1983d8f1678cf6dae3be41458d2ad5ef5379765741d807a5403a1ac526a8c1e\": container with ID starting with f1983d8f1678cf6dae3be41458d2ad5ef5379765741d807a5403a1ac526a8c1e not found: ID does not exist" Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.740647 4696 scope.go:117] "RemoveContainer" containerID="ffb65f0f246951fcac164ce9a2886d85e3d193155e6a753292814e4af9f7657a" Mar 21 09:04:14 crc kubenswrapper[4696]: E0321 09:04:14.741116 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffb65f0f246951fcac164ce9a2886d85e3d193155e6a753292814e4af9f7657a\": container with ID starting with ffb65f0f246951fcac164ce9a2886d85e3d193155e6a753292814e4af9f7657a not found: ID does not exist" containerID="ffb65f0f246951fcac164ce9a2886d85e3d193155e6a753292814e4af9f7657a" Mar 21 09:04:14 crc kubenswrapper[4696]: I0321 09:04:14.741153 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffb65f0f246951fcac164ce9a2886d85e3d193155e6a753292814e4af9f7657a"} err="failed to get container status \"ffb65f0f246951fcac164ce9a2886d85e3d193155e6a753292814e4af9f7657a\": rpc error: code = NotFound desc = could not find container \"ffb65f0f246951fcac164ce9a2886d85e3d193155e6a753292814e4af9f7657a\": container with ID starting with ffb65f0f246951fcac164ce9a2886d85e3d193155e6a753292814e4af9f7657a not found: ID does not exist" Mar 21 09:04:16 crc kubenswrapper[4696]: I0321 09:04:16.546062 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fc4435a-7193-4c62-99c4-008ac4555225" path="/var/lib/kubelet/pods/7fc4435a-7193-4c62-99c4-008ac4555225/volumes" Mar 21 09:05:01 crc kubenswrapper[4696]: I0321 09:05:01.594520 4696 scope.go:117] "RemoveContainer" containerID="fcc17446e6d74173152ee59e6caa9c2656bf271716767e63d5662f0be42c52a3" Mar 21 09:05:25 crc kubenswrapper[4696]: I0321 09:05:25.698786 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pw6kc"] Mar 21 09:05:25 crc kubenswrapper[4696]: E0321 09:05:25.699696 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d52bfab3-d526-433b-86ac-3213f08251d6" containerName="oc" Mar 21 09:05:25 crc kubenswrapper[4696]: I0321 09:05:25.699710 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d52bfab3-d526-433b-86ac-3213f08251d6" containerName="oc" Mar 21 09:05:25 crc kubenswrapper[4696]: E0321 09:05:25.699730 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc4435a-7193-4c62-99c4-008ac4555225" containerName="extract-content" Mar 21 09:05:25 crc kubenswrapper[4696]: I0321 09:05:25.699736 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc4435a-7193-4c62-99c4-008ac4555225" containerName="extract-content" Mar 21 09:05:25 crc kubenswrapper[4696]: E0321 09:05:25.699747 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc4435a-7193-4c62-99c4-008ac4555225" containerName="extract-utilities" Mar 21 09:05:25 crc kubenswrapper[4696]: I0321 09:05:25.699754 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc4435a-7193-4c62-99c4-008ac4555225" containerName="extract-utilities" Mar 21 09:05:25 crc kubenswrapper[4696]: E0321 09:05:25.699779 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc4435a-7193-4c62-99c4-008ac4555225" containerName="registry-server" Mar 21 09:05:25 crc kubenswrapper[4696]: I0321 09:05:25.699789 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc4435a-7193-4c62-99c4-008ac4555225" containerName="registry-server" Mar 21 09:05:25 crc kubenswrapper[4696]: I0321 09:05:25.700021 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="d52bfab3-d526-433b-86ac-3213f08251d6" containerName="oc" Mar 21 09:05:25 crc kubenswrapper[4696]: I0321 09:05:25.700041 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fc4435a-7193-4c62-99c4-008ac4555225" containerName="registry-server" Mar 21 09:05:25 crc kubenswrapper[4696]: I0321 09:05:25.701559 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pw6kc" Mar 21 09:05:25 crc kubenswrapper[4696]: I0321 09:05:25.712241 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pw6kc"] Mar 21 09:05:25 crc kubenswrapper[4696]: I0321 09:05:25.793966 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b12d706a-e648-4c5a-988e-a46f641bee27-utilities\") pod \"redhat-marketplace-pw6kc\" (UID: \"b12d706a-e648-4c5a-988e-a46f641bee27\") " pod="openshift-marketplace/redhat-marketplace-pw6kc" Mar 21 09:05:25 crc kubenswrapper[4696]: I0321 09:05:25.794299 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b12d706a-e648-4c5a-988e-a46f641bee27-catalog-content\") pod \"redhat-marketplace-pw6kc\" (UID: \"b12d706a-e648-4c5a-988e-a46f641bee27\") " pod="openshift-marketplace/redhat-marketplace-pw6kc" Mar 21 09:05:25 crc kubenswrapper[4696]: I0321 09:05:25.794378 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxpd6\" (UniqueName: \"kubernetes.io/projected/b12d706a-e648-4c5a-988e-a46f641bee27-kube-api-access-nxpd6\") pod \"redhat-marketplace-pw6kc\" (UID: \"b12d706a-e648-4c5a-988e-a46f641bee27\") " pod="openshift-marketplace/redhat-marketplace-pw6kc" Mar 21 09:05:25 crc kubenswrapper[4696]: I0321 09:05:25.896006 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxpd6\" (UniqueName: \"kubernetes.io/projected/b12d706a-e648-4c5a-988e-a46f641bee27-kube-api-access-nxpd6\") pod \"redhat-marketplace-pw6kc\" (UID: \"b12d706a-e648-4c5a-988e-a46f641bee27\") " pod="openshift-marketplace/redhat-marketplace-pw6kc" Mar 21 09:05:25 crc kubenswrapper[4696]: I0321 09:05:25.896222 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b12d706a-e648-4c5a-988e-a46f641bee27-utilities\") pod \"redhat-marketplace-pw6kc\" (UID: \"b12d706a-e648-4c5a-988e-a46f641bee27\") " pod="openshift-marketplace/redhat-marketplace-pw6kc" Mar 21 09:05:25 crc kubenswrapper[4696]: I0321 09:05:25.896260 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b12d706a-e648-4c5a-988e-a46f641bee27-catalog-content\") pod \"redhat-marketplace-pw6kc\" (UID: \"b12d706a-e648-4c5a-988e-a46f641bee27\") " pod="openshift-marketplace/redhat-marketplace-pw6kc" Mar 21 09:05:25 crc kubenswrapper[4696]: I0321 09:05:25.896786 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b12d706a-e648-4c5a-988e-a46f641bee27-catalog-content\") pod \"redhat-marketplace-pw6kc\" (UID: \"b12d706a-e648-4c5a-988e-a46f641bee27\") " pod="openshift-marketplace/redhat-marketplace-pw6kc" Mar 21 09:05:25 crc kubenswrapper[4696]: I0321 09:05:25.896935 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b12d706a-e648-4c5a-988e-a46f641bee27-utilities\") pod \"redhat-marketplace-pw6kc\" (UID: \"b12d706a-e648-4c5a-988e-a46f641bee27\") " pod="openshift-marketplace/redhat-marketplace-pw6kc" Mar 21 09:05:25 crc kubenswrapper[4696]: I0321 09:05:25.918571 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxpd6\" (UniqueName: \"kubernetes.io/projected/b12d706a-e648-4c5a-988e-a46f641bee27-kube-api-access-nxpd6\") pod \"redhat-marketplace-pw6kc\" (UID: \"b12d706a-e648-4c5a-988e-a46f641bee27\") " pod="openshift-marketplace/redhat-marketplace-pw6kc" Mar 21 09:05:26 crc kubenswrapper[4696]: I0321 09:05:26.024896 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pw6kc" Mar 21 09:05:26 crc kubenswrapper[4696]: I0321 09:05:26.525780 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pw6kc"] Mar 21 09:05:26 crc kubenswrapper[4696]: W0321 09:05:26.528180 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb12d706a_e648_4c5a_988e_a46f641bee27.slice/crio-69ba90360c99ad38a42bf153b6d5f75cf589468615aa069fc5236bf59272333b WatchSource:0}: Error finding container 69ba90360c99ad38a42bf153b6d5f75cf589468615aa069fc5236bf59272333b: Status 404 returned error can't find the container with id 69ba90360c99ad38a42bf153b6d5f75cf589468615aa069fc5236bf59272333b Mar 21 09:05:27 crc kubenswrapper[4696]: I0321 09:05:27.311848 4696 generic.go:334] "Generic (PLEG): container finished" podID="b12d706a-e648-4c5a-988e-a46f641bee27" containerID="da6408abba52852c2a679fa32ad962a7a0d6e20d5d012edb04c20691a9e2b771" exitCode=0 Mar 21 09:05:27 crc kubenswrapper[4696]: I0321 09:05:27.312043 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pw6kc" event={"ID":"b12d706a-e648-4c5a-988e-a46f641bee27","Type":"ContainerDied","Data":"da6408abba52852c2a679fa32ad962a7a0d6e20d5d012edb04c20691a9e2b771"} Mar 21 09:05:27 crc kubenswrapper[4696]: I0321 09:05:27.312119 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pw6kc" event={"ID":"b12d706a-e648-4c5a-988e-a46f641bee27","Type":"ContainerStarted","Data":"69ba90360c99ad38a42bf153b6d5f75cf589468615aa069fc5236bf59272333b"} Mar 21 09:05:28 crc kubenswrapper[4696]: I0321 09:05:28.322735 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pw6kc" event={"ID":"b12d706a-e648-4c5a-988e-a46f641bee27","Type":"ContainerStarted","Data":"1fbbb141f66683d8ae3f1edda1b602bfba3c089d8c3d97886bd59c7660802c07"} Mar 21 09:05:29 crc kubenswrapper[4696]: I0321 09:05:29.334873 4696 generic.go:334] "Generic (PLEG): container finished" podID="b12d706a-e648-4c5a-988e-a46f641bee27" containerID="1fbbb141f66683d8ae3f1edda1b602bfba3c089d8c3d97886bd59c7660802c07" exitCode=0 Mar 21 09:05:29 crc kubenswrapper[4696]: I0321 09:05:29.334928 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pw6kc" event={"ID":"b12d706a-e648-4c5a-988e-a46f641bee27","Type":"ContainerDied","Data":"1fbbb141f66683d8ae3f1edda1b602bfba3c089d8c3d97886bd59c7660802c07"} Mar 21 09:05:30 crc kubenswrapper[4696]: I0321 09:05:30.340988 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:05:30 crc kubenswrapper[4696]: I0321 09:05:30.341919 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:05:30 crc kubenswrapper[4696]: I0321 09:05:30.346286 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pw6kc" event={"ID":"b12d706a-e648-4c5a-988e-a46f641bee27","Type":"ContainerStarted","Data":"b268826df26d0d69fcbccd73ae9d00c409c61671851571128b6c32eecef70f22"} Mar 21 09:05:30 crc kubenswrapper[4696]: I0321 09:05:30.383389 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pw6kc" podStartSLOduration=2.866763148 podStartE2EDuration="5.383361255s" podCreationTimestamp="2026-03-21 09:05:25 +0000 UTC" firstStartedPulling="2026-03-21 09:05:27.3139982 +0000 UTC m=+2261.434878913" lastFinishedPulling="2026-03-21 09:05:29.830596307 +0000 UTC m=+2263.951477020" observedRunningTime="2026-03-21 09:05:30.36471234 +0000 UTC m=+2264.485593063" watchObservedRunningTime="2026-03-21 09:05:30.383361255 +0000 UTC m=+2264.504242008" Mar 21 09:05:36 crc kubenswrapper[4696]: I0321 09:05:36.025945 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pw6kc" Mar 21 09:05:36 crc kubenswrapper[4696]: I0321 09:05:36.026661 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pw6kc" Mar 21 09:05:36 crc kubenswrapper[4696]: I0321 09:05:36.079164 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pw6kc" Mar 21 09:05:36 crc kubenswrapper[4696]: I0321 09:05:36.437925 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pw6kc" Mar 21 09:05:37 crc kubenswrapper[4696]: I0321 09:05:37.486843 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pw6kc"] Mar 21 09:05:38 crc kubenswrapper[4696]: I0321 09:05:38.420899 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pw6kc" podUID="b12d706a-e648-4c5a-988e-a46f641bee27" containerName="registry-server" containerID="cri-o://b268826df26d0d69fcbccd73ae9d00c409c61671851571128b6c32eecef70f22" gracePeriod=2 Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.428073 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pw6kc" Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.431691 4696 generic.go:334] "Generic (PLEG): container finished" podID="b12d706a-e648-4c5a-988e-a46f641bee27" containerID="b268826df26d0d69fcbccd73ae9d00c409c61671851571128b6c32eecef70f22" exitCode=0 Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.431735 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pw6kc" event={"ID":"b12d706a-e648-4c5a-988e-a46f641bee27","Type":"ContainerDied","Data":"b268826df26d0d69fcbccd73ae9d00c409c61671851571128b6c32eecef70f22"} Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.431764 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pw6kc" event={"ID":"b12d706a-e648-4c5a-988e-a46f641bee27","Type":"ContainerDied","Data":"69ba90360c99ad38a42bf153b6d5f75cf589468615aa069fc5236bf59272333b"} Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.431785 4696 scope.go:117] "RemoveContainer" containerID="b268826df26d0d69fcbccd73ae9d00c409c61671851571128b6c32eecef70f22" Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.431995 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pw6kc" Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.461120 4696 scope.go:117] "RemoveContainer" containerID="1fbbb141f66683d8ae3f1edda1b602bfba3c089d8c3d97886bd59c7660802c07" Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.486073 4696 scope.go:117] "RemoveContainer" containerID="da6408abba52852c2a679fa32ad962a7a0d6e20d5d012edb04c20691a9e2b771" Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.523938 4696 scope.go:117] "RemoveContainer" containerID="b268826df26d0d69fcbccd73ae9d00c409c61671851571128b6c32eecef70f22" Mar 21 09:05:39 crc kubenswrapper[4696]: E0321 09:05:39.524248 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b268826df26d0d69fcbccd73ae9d00c409c61671851571128b6c32eecef70f22\": container with ID starting with b268826df26d0d69fcbccd73ae9d00c409c61671851571128b6c32eecef70f22 not found: ID does not exist" containerID="b268826df26d0d69fcbccd73ae9d00c409c61671851571128b6c32eecef70f22" Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.524270 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b268826df26d0d69fcbccd73ae9d00c409c61671851571128b6c32eecef70f22"} err="failed to get container status \"b268826df26d0d69fcbccd73ae9d00c409c61671851571128b6c32eecef70f22\": rpc error: code = NotFound desc = could not find container \"b268826df26d0d69fcbccd73ae9d00c409c61671851571128b6c32eecef70f22\": container with ID starting with b268826df26d0d69fcbccd73ae9d00c409c61671851571128b6c32eecef70f22 not found: ID does not exist" Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.524290 4696 scope.go:117] "RemoveContainer" containerID="1fbbb141f66683d8ae3f1edda1b602bfba3c089d8c3d97886bd59c7660802c07" Mar 21 09:05:39 crc kubenswrapper[4696]: E0321 09:05:39.525465 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fbbb141f66683d8ae3f1edda1b602bfba3c089d8c3d97886bd59c7660802c07\": container with ID starting with 1fbbb141f66683d8ae3f1edda1b602bfba3c089d8c3d97886bd59c7660802c07 not found: ID does not exist" containerID="1fbbb141f66683d8ae3f1edda1b602bfba3c089d8c3d97886bd59c7660802c07" Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.525490 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fbbb141f66683d8ae3f1edda1b602bfba3c089d8c3d97886bd59c7660802c07"} err="failed to get container status \"1fbbb141f66683d8ae3f1edda1b602bfba3c089d8c3d97886bd59c7660802c07\": rpc error: code = NotFound desc = could not find container \"1fbbb141f66683d8ae3f1edda1b602bfba3c089d8c3d97886bd59c7660802c07\": container with ID starting with 1fbbb141f66683d8ae3f1edda1b602bfba3c089d8c3d97886bd59c7660802c07 not found: ID does not exist" Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.525508 4696 scope.go:117] "RemoveContainer" containerID="da6408abba52852c2a679fa32ad962a7a0d6e20d5d012edb04c20691a9e2b771" Mar 21 09:05:39 crc kubenswrapper[4696]: E0321 09:05:39.525784 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da6408abba52852c2a679fa32ad962a7a0d6e20d5d012edb04c20691a9e2b771\": container with ID starting with da6408abba52852c2a679fa32ad962a7a0d6e20d5d012edb04c20691a9e2b771 not found: ID does not exist" containerID="da6408abba52852c2a679fa32ad962a7a0d6e20d5d012edb04c20691a9e2b771" Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.525804 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da6408abba52852c2a679fa32ad962a7a0d6e20d5d012edb04c20691a9e2b771"} err="failed to get container status \"da6408abba52852c2a679fa32ad962a7a0d6e20d5d012edb04c20691a9e2b771\": rpc error: code = NotFound desc = could not find container \"da6408abba52852c2a679fa32ad962a7a0d6e20d5d012edb04c20691a9e2b771\": container with ID starting with da6408abba52852c2a679fa32ad962a7a0d6e20d5d012edb04c20691a9e2b771 not found: ID does not exist" Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.530653 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxpd6\" (UniqueName: \"kubernetes.io/projected/b12d706a-e648-4c5a-988e-a46f641bee27-kube-api-access-nxpd6\") pod \"b12d706a-e648-4c5a-988e-a46f641bee27\" (UID: \"b12d706a-e648-4c5a-988e-a46f641bee27\") " Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.530686 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b12d706a-e648-4c5a-988e-a46f641bee27-utilities\") pod \"b12d706a-e648-4c5a-988e-a46f641bee27\" (UID: \"b12d706a-e648-4c5a-988e-a46f641bee27\") " Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.530727 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b12d706a-e648-4c5a-988e-a46f641bee27-catalog-content\") pod \"b12d706a-e648-4c5a-988e-a46f641bee27\" (UID: \"b12d706a-e648-4c5a-988e-a46f641bee27\") " Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.531493 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b12d706a-e648-4c5a-988e-a46f641bee27-utilities" (OuterVolumeSpecName: "utilities") pod "b12d706a-e648-4c5a-988e-a46f641bee27" (UID: "b12d706a-e648-4c5a-988e-a46f641bee27"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.536723 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b12d706a-e648-4c5a-988e-a46f641bee27-kube-api-access-nxpd6" (OuterVolumeSpecName: "kube-api-access-nxpd6") pod "b12d706a-e648-4c5a-988e-a46f641bee27" (UID: "b12d706a-e648-4c5a-988e-a46f641bee27"). InnerVolumeSpecName "kube-api-access-nxpd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.556603 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b12d706a-e648-4c5a-988e-a46f641bee27-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b12d706a-e648-4c5a-988e-a46f641bee27" (UID: "b12d706a-e648-4c5a-988e-a46f641bee27"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.633686 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxpd6\" (UniqueName: \"kubernetes.io/projected/b12d706a-e648-4c5a-988e-a46f641bee27-kube-api-access-nxpd6\") on node \"crc\" DevicePath \"\"" Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.633720 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b12d706a-e648-4c5a-988e-a46f641bee27-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.633736 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b12d706a-e648-4c5a-988e-a46f641bee27-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.765689 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pw6kc"] Mar 21 09:05:39 crc kubenswrapper[4696]: I0321 09:05:39.774657 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pw6kc"] Mar 21 09:05:40 crc kubenswrapper[4696]: I0321 09:05:40.546347 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b12d706a-e648-4c5a-988e-a46f641bee27" path="/var/lib/kubelet/pods/b12d706a-e648-4c5a-988e-a46f641bee27/volumes" Mar 21 09:06:00 crc kubenswrapper[4696]: I0321 09:06:00.151141 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568066-8qgbv"] Mar 21 09:06:00 crc kubenswrapper[4696]: E0321 09:06:00.152198 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12d706a-e648-4c5a-988e-a46f641bee27" containerName="registry-server" Mar 21 09:06:00 crc kubenswrapper[4696]: I0321 09:06:00.152215 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12d706a-e648-4c5a-988e-a46f641bee27" containerName="registry-server" Mar 21 09:06:00 crc kubenswrapper[4696]: E0321 09:06:00.152242 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12d706a-e648-4c5a-988e-a46f641bee27" containerName="extract-utilities" Mar 21 09:06:00 crc kubenswrapper[4696]: I0321 09:06:00.152250 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12d706a-e648-4c5a-988e-a46f641bee27" containerName="extract-utilities" Mar 21 09:06:00 crc kubenswrapper[4696]: E0321 09:06:00.152257 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12d706a-e648-4c5a-988e-a46f641bee27" containerName="extract-content" Mar 21 09:06:00 crc kubenswrapper[4696]: I0321 09:06:00.152264 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12d706a-e648-4c5a-988e-a46f641bee27" containerName="extract-content" Mar 21 09:06:00 crc kubenswrapper[4696]: I0321 09:06:00.152519 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="b12d706a-e648-4c5a-988e-a46f641bee27" containerName="registry-server" Mar 21 09:06:00 crc kubenswrapper[4696]: I0321 09:06:00.153319 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568066-8qgbv" Mar 21 09:06:00 crc kubenswrapper[4696]: I0321 09:06:00.155728 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:06:00 crc kubenswrapper[4696]: I0321 09:06:00.155858 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:06:00 crc kubenswrapper[4696]: I0321 09:06:00.155996 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:06:00 crc kubenswrapper[4696]: I0321 09:06:00.177932 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568066-8qgbv"] Mar 21 09:06:00 crc kubenswrapper[4696]: I0321 09:06:00.215462 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gt8s\" (UniqueName: \"kubernetes.io/projected/e5721fef-e6a4-4e01-a722-5436a7ce7eff-kube-api-access-9gt8s\") pod \"auto-csr-approver-29568066-8qgbv\" (UID: \"e5721fef-e6a4-4e01-a722-5436a7ce7eff\") " pod="openshift-infra/auto-csr-approver-29568066-8qgbv" Mar 21 09:06:00 crc kubenswrapper[4696]: I0321 09:06:00.318204 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gt8s\" (UniqueName: \"kubernetes.io/projected/e5721fef-e6a4-4e01-a722-5436a7ce7eff-kube-api-access-9gt8s\") pod \"auto-csr-approver-29568066-8qgbv\" (UID: \"e5721fef-e6a4-4e01-a722-5436a7ce7eff\") " pod="openshift-infra/auto-csr-approver-29568066-8qgbv" Mar 21 09:06:00 crc kubenswrapper[4696]: I0321 09:06:00.336719 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gt8s\" (UniqueName: \"kubernetes.io/projected/e5721fef-e6a4-4e01-a722-5436a7ce7eff-kube-api-access-9gt8s\") pod \"auto-csr-approver-29568066-8qgbv\" (UID: \"e5721fef-e6a4-4e01-a722-5436a7ce7eff\") " pod="openshift-infra/auto-csr-approver-29568066-8qgbv" Mar 21 09:06:00 crc kubenswrapper[4696]: I0321 09:06:00.341039 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:06:00 crc kubenswrapper[4696]: I0321 09:06:00.341103 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:06:00 crc kubenswrapper[4696]: I0321 09:06:00.478884 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568066-8qgbv" Mar 21 09:06:00 crc kubenswrapper[4696]: I0321 09:06:00.908883 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568066-8qgbv"] Mar 21 09:06:00 crc kubenswrapper[4696]: W0321 09:06:00.927167 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5721fef_e6a4_4e01_a722_5436a7ce7eff.slice/crio-b1a4faee944cddb44a4782dd6ea53f19ab53eb43869d1d863d352502b2979711 WatchSource:0}: Error finding container b1a4faee944cddb44a4782dd6ea53f19ab53eb43869d1d863d352502b2979711: Status 404 returned error can't find the container with id b1a4faee944cddb44a4782dd6ea53f19ab53eb43869d1d863d352502b2979711 Mar 21 09:06:01 crc kubenswrapper[4696]: I0321 09:06:01.626668 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568066-8qgbv" event={"ID":"e5721fef-e6a4-4e01-a722-5436a7ce7eff","Type":"ContainerStarted","Data":"b1a4faee944cddb44a4782dd6ea53f19ab53eb43869d1d863d352502b2979711"} Mar 21 09:06:02 crc kubenswrapper[4696]: I0321 09:06:02.638779 4696 generic.go:334] "Generic (PLEG): container finished" podID="e5721fef-e6a4-4e01-a722-5436a7ce7eff" containerID="71c0fab430ec4d9914d39524313c6dbe3ae1f7aa18064eb07d884b0d68b6f3e7" exitCode=0 Mar 21 09:06:02 crc kubenswrapper[4696]: I0321 09:06:02.638930 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568066-8qgbv" event={"ID":"e5721fef-e6a4-4e01-a722-5436a7ce7eff","Type":"ContainerDied","Data":"71c0fab430ec4d9914d39524313c6dbe3ae1f7aa18064eb07d884b0d68b6f3e7"} Mar 21 09:06:04 crc kubenswrapper[4696]: I0321 09:06:04.024484 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568066-8qgbv" Mar 21 09:06:04 crc kubenswrapper[4696]: I0321 09:06:04.101442 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gt8s\" (UniqueName: \"kubernetes.io/projected/e5721fef-e6a4-4e01-a722-5436a7ce7eff-kube-api-access-9gt8s\") pod \"e5721fef-e6a4-4e01-a722-5436a7ce7eff\" (UID: \"e5721fef-e6a4-4e01-a722-5436a7ce7eff\") " Mar 21 09:06:04 crc kubenswrapper[4696]: I0321 09:06:04.107809 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5721fef-e6a4-4e01-a722-5436a7ce7eff-kube-api-access-9gt8s" (OuterVolumeSpecName: "kube-api-access-9gt8s") pod "e5721fef-e6a4-4e01-a722-5436a7ce7eff" (UID: "e5721fef-e6a4-4e01-a722-5436a7ce7eff"). InnerVolumeSpecName "kube-api-access-9gt8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:06:04 crc kubenswrapper[4696]: I0321 09:06:04.204961 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gt8s\" (UniqueName: \"kubernetes.io/projected/e5721fef-e6a4-4e01-a722-5436a7ce7eff-kube-api-access-9gt8s\") on node \"crc\" DevicePath \"\"" Mar 21 09:06:04 crc kubenswrapper[4696]: I0321 09:06:04.663216 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568066-8qgbv" event={"ID":"e5721fef-e6a4-4e01-a722-5436a7ce7eff","Type":"ContainerDied","Data":"b1a4faee944cddb44a4782dd6ea53f19ab53eb43869d1d863d352502b2979711"} Mar 21 09:06:04 crc kubenswrapper[4696]: I0321 09:06:04.663267 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1a4faee944cddb44a4782dd6ea53f19ab53eb43869d1d863d352502b2979711" Mar 21 09:06:04 crc kubenswrapper[4696]: I0321 09:06:04.663328 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568066-8qgbv" Mar 21 09:06:05 crc kubenswrapper[4696]: I0321 09:06:05.094596 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568060-qbn8d"] Mar 21 09:06:05 crc kubenswrapper[4696]: I0321 09:06:05.103747 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568060-qbn8d"] Mar 21 09:06:06 crc kubenswrapper[4696]: I0321 09:06:06.545470 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="140b0556-76e9-4138-9395-3c7bacf16cb2" path="/var/lib/kubelet/pods/140b0556-76e9-4138-9395-3c7bacf16cb2/volumes" Mar 21 09:06:30 crc kubenswrapper[4696]: I0321 09:06:30.341374 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:06:30 crc kubenswrapper[4696]: I0321 09:06:30.341990 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:06:30 crc kubenswrapper[4696]: I0321 09:06:30.342039 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 09:06:30 crc kubenswrapper[4696]: I0321 09:06:30.342890 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263"} pod="openshift-machine-config-operator/machine-config-daemon-z7srw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 09:06:30 crc kubenswrapper[4696]: I0321 09:06:30.342944 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" containerID="cri-o://d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" gracePeriod=600 Mar 21 09:06:30 crc kubenswrapper[4696]: E0321 09:06:30.482330 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:06:30 crc kubenswrapper[4696]: I0321 09:06:30.922631 4696 generic.go:334] "Generic (PLEG): container finished" podID="daaf227a-2305-495e-8495-a280abcd8e10" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" exitCode=0 Mar 21 09:06:30 crc kubenswrapper[4696]: I0321 09:06:30.922677 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerDied","Data":"d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263"} Mar 21 09:06:30 crc kubenswrapper[4696]: I0321 09:06:30.922709 4696 scope.go:117] "RemoveContainer" containerID="3187c3e490f50d1673fbeb2f54747e27d740a56d231e0d24142915c7db238efa" Mar 21 09:06:30 crc kubenswrapper[4696]: I0321 09:06:30.923506 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:06:30 crc kubenswrapper[4696]: E0321 09:06:30.923885 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:06:41 crc kubenswrapper[4696]: I0321 09:06:41.534778 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:06:41 crc kubenswrapper[4696]: E0321 09:06:41.535483 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:06:53 crc kubenswrapper[4696]: I0321 09:06:53.535316 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:06:53 crc kubenswrapper[4696]: E0321 09:06:53.536127 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:07:01 crc kubenswrapper[4696]: I0321 09:07:01.737531 4696 scope.go:117] "RemoveContainer" containerID="8409d25fad0117a45a1a1c8ba21857376c1508984d7527330e44a7c31208f36e" Mar 21 09:07:04 crc kubenswrapper[4696]: I0321 09:07:04.257445 4696 generic.go:334] "Generic (PLEG): container finished" podID="50e8cc4d-e8bf-4f1a-9c84-f40af2392cde" containerID="fb47bfe25466eaaa65a67be0c30bc309e37f7c418327b2d31077b82071a5b6d0" exitCode=0 Mar 21 09:07:04 crc kubenswrapper[4696]: I0321 09:07:04.257540 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" event={"ID":"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde","Type":"ContainerDied","Data":"fb47bfe25466eaaa65a67be0c30bc309e37f7c418327b2d31077b82071a5b6d0"} Mar 21 09:07:04 crc kubenswrapper[4696]: I0321 09:07:04.534953 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:07:04 crc kubenswrapper[4696]: E0321 09:07:04.535239 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:07:05 crc kubenswrapper[4696]: I0321 09:07:05.704960 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" Mar 21 09:07:05 crc kubenswrapper[4696]: I0321 09:07:05.758622 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-libvirt-combined-ca-bundle\") pod \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\" (UID: \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\") " Mar 21 09:07:05 crc kubenswrapper[4696]: I0321 09:07:05.758773 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-inventory\") pod \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\" (UID: \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\") " Mar 21 09:07:05 crc kubenswrapper[4696]: I0321 09:07:05.758806 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-ssh-key-openstack-edpm-ipam\") pod \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\" (UID: \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\") " Mar 21 09:07:05 crc kubenswrapper[4696]: I0321 09:07:05.759054 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-libvirt-secret-0\") pod \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\" (UID: \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\") " Mar 21 09:07:05 crc kubenswrapper[4696]: I0321 09:07:05.759090 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrj5v\" (UniqueName: \"kubernetes.io/projected/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-kube-api-access-rrj5v\") pod \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\" (UID: \"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde\") " Mar 21 09:07:05 crc kubenswrapper[4696]: I0321 09:07:05.766579 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-kube-api-access-rrj5v" (OuterVolumeSpecName: "kube-api-access-rrj5v") pod "50e8cc4d-e8bf-4f1a-9c84-f40af2392cde" (UID: "50e8cc4d-e8bf-4f1a-9c84-f40af2392cde"). InnerVolumeSpecName "kube-api-access-rrj5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:07:05 crc kubenswrapper[4696]: I0321 09:07:05.766621 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "50e8cc4d-e8bf-4f1a-9c84-f40af2392cde" (UID: "50e8cc4d-e8bf-4f1a-9c84-f40af2392cde"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:07:05 crc kubenswrapper[4696]: I0321 09:07:05.788235 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-inventory" (OuterVolumeSpecName: "inventory") pod "50e8cc4d-e8bf-4f1a-9c84-f40af2392cde" (UID: "50e8cc4d-e8bf-4f1a-9c84-f40af2392cde"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:07:05 crc kubenswrapper[4696]: I0321 09:07:05.789148 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "50e8cc4d-e8bf-4f1a-9c84-f40af2392cde" (UID: "50e8cc4d-e8bf-4f1a-9c84-f40af2392cde"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:07:05 crc kubenswrapper[4696]: I0321 09:07:05.800429 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "50e8cc4d-e8bf-4f1a-9c84-f40af2392cde" (UID: "50e8cc4d-e8bf-4f1a-9c84-f40af2392cde"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:07:05 crc kubenswrapper[4696]: I0321 09:07:05.861615 4696 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Mar 21 09:07:05 crc kubenswrapper[4696]: I0321 09:07:05.861659 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrj5v\" (UniqueName: \"kubernetes.io/projected/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-kube-api-access-rrj5v\") on node \"crc\" DevicePath \"\"" Mar 21 09:07:05 crc kubenswrapper[4696]: I0321 09:07:05.861675 4696 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 09:07:05 crc kubenswrapper[4696]: I0321 09:07:05.861689 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 09:07:05 crc kubenswrapper[4696]: I0321 09:07:05.861700 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50e8cc4d-e8bf-4f1a-9c84-f40af2392cde-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.278146 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" event={"ID":"50e8cc4d-e8bf-4f1a-9c84-f40af2392cde","Type":"ContainerDied","Data":"0e305f921e2b2b2aa6bba2e1428e4e067e0b0b8ca3ff53575edabe788eee8b96"} Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.278430 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e305f921e2b2b2aa6bba2e1428e4e067e0b0b8ca3ff53575edabe788eee8b96" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.278385 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.388405 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl"] Mar 21 09:07:06 crc kubenswrapper[4696]: E0321 09:07:06.389474 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50e8cc4d-e8bf-4f1a-9c84-f40af2392cde" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.389496 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="50e8cc4d-e8bf-4f1a-9c84-f40af2392cde" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 21 09:07:06 crc kubenswrapper[4696]: E0321 09:07:06.389512 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5721fef-e6a4-4e01-a722-5436a7ce7eff" containerName="oc" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.389520 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5721fef-e6a4-4e01-a722-5436a7ce7eff" containerName="oc" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.389939 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="50e8cc4d-e8bf-4f1a-9c84-f40af2392cde" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.389968 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5721fef-e6a4-4e01-a722-5436a7ce7eff" containerName="oc" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.393066 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.399271 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bqp9r" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.399682 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.399681 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.399936 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.400077 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.400143 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.400330 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.412135 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl"] Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.478124 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.478190 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.478230 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.478394 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.478450 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbpvt\" (UniqueName: \"kubernetes.io/projected/14ded188-15ec-4d0c-a462-d1160dfb5b47-kube-api-access-zbpvt\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.478518 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.478772 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.478801 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.478840 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.478870 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.479199 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.581241 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.581394 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.581416 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.581438 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.581463 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.581491 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.581539 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.581585 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.581620 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.581647 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.581685 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbpvt\" (UniqueName: \"kubernetes.io/projected/14ded188-15ec-4d0c-a462-d1160dfb5b47-kube-api-access-zbpvt\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.582861 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.585853 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.586563 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.589604 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.589710 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.589873 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.589935 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.590060 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.590148 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.590088 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.599624 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbpvt\" (UniqueName: \"kubernetes.io/projected/14ded188-15ec-4d0c-a462-d1160dfb5b47-kube-api-access-zbpvt\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dpstl\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:06 crc kubenswrapper[4696]: I0321 09:07:06.729034 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:07:07 crc kubenswrapper[4696]: I0321 09:07:07.248211 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl"] Mar 21 09:07:07 crc kubenswrapper[4696]: I0321 09:07:07.289682 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" event={"ID":"14ded188-15ec-4d0c-a462-d1160dfb5b47","Type":"ContainerStarted","Data":"e0c69d94c374ed50a79270341c43e3ae98d4cde4c67d43bb9af12641cb3ecefb"} Mar 21 09:07:08 crc kubenswrapper[4696]: I0321 09:07:08.300509 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" event={"ID":"14ded188-15ec-4d0c-a462-d1160dfb5b47","Type":"ContainerStarted","Data":"ec720d24aa26ccc83152227d846ea6c91cb778773268fefee8acac4e5fb8230f"} Mar 21 09:07:08 crc kubenswrapper[4696]: I0321 09:07:08.321484 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" podStartSLOduration=1.8840497489999999 podStartE2EDuration="2.321466379s" podCreationTimestamp="2026-03-21 09:07:06 +0000 UTC" firstStartedPulling="2026-03-21 09:07:07.244559833 +0000 UTC m=+2361.365440536" lastFinishedPulling="2026-03-21 09:07:07.681976453 +0000 UTC m=+2361.802857166" observedRunningTime="2026-03-21 09:07:08.317410887 +0000 UTC m=+2362.438291600" watchObservedRunningTime="2026-03-21 09:07:08.321466379 +0000 UTC m=+2362.442347102" Mar 21 09:07:18 crc kubenswrapper[4696]: I0321 09:07:18.535267 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:07:18 crc kubenswrapper[4696]: E0321 09:07:18.536052 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:07:32 crc kubenswrapper[4696]: I0321 09:07:32.535582 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:07:32 crc kubenswrapper[4696]: E0321 09:07:32.536550 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:07:47 crc kubenswrapper[4696]: I0321 09:07:47.535340 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:07:47 crc kubenswrapper[4696]: E0321 09:07:47.536186 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:08:00 crc kubenswrapper[4696]: I0321 09:08:00.150262 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568068-6f9cd"] Mar 21 09:08:00 crc kubenswrapper[4696]: I0321 09:08:00.152463 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568068-6f9cd" Mar 21 09:08:00 crc kubenswrapper[4696]: I0321 09:08:00.154543 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:08:00 crc kubenswrapper[4696]: I0321 09:08:00.154605 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:08:00 crc kubenswrapper[4696]: I0321 09:08:00.155355 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:08:00 crc kubenswrapper[4696]: I0321 09:08:00.167570 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568068-6f9cd"] Mar 21 09:08:00 crc kubenswrapper[4696]: I0321 09:08:00.188350 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x92fs\" (UniqueName: \"kubernetes.io/projected/1bae18e6-4af5-467d-a736-78b4d04a0fa3-kube-api-access-x92fs\") pod \"auto-csr-approver-29568068-6f9cd\" (UID: \"1bae18e6-4af5-467d-a736-78b4d04a0fa3\") " pod="openshift-infra/auto-csr-approver-29568068-6f9cd" Mar 21 09:08:00 crc kubenswrapper[4696]: I0321 09:08:00.290711 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x92fs\" (UniqueName: \"kubernetes.io/projected/1bae18e6-4af5-467d-a736-78b4d04a0fa3-kube-api-access-x92fs\") pod \"auto-csr-approver-29568068-6f9cd\" (UID: \"1bae18e6-4af5-467d-a736-78b4d04a0fa3\") " pod="openshift-infra/auto-csr-approver-29568068-6f9cd" Mar 21 09:08:00 crc kubenswrapper[4696]: I0321 09:08:00.310255 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x92fs\" (UniqueName: \"kubernetes.io/projected/1bae18e6-4af5-467d-a736-78b4d04a0fa3-kube-api-access-x92fs\") pod \"auto-csr-approver-29568068-6f9cd\" (UID: \"1bae18e6-4af5-467d-a736-78b4d04a0fa3\") " pod="openshift-infra/auto-csr-approver-29568068-6f9cd" Mar 21 09:08:00 crc kubenswrapper[4696]: I0321 09:08:00.476205 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568068-6f9cd" Mar 21 09:08:01 crc kubenswrapper[4696]: I0321 09:08:01.006777 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 09:08:01 crc kubenswrapper[4696]: I0321 09:08:01.027113 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568068-6f9cd"] Mar 21 09:08:01 crc kubenswrapper[4696]: I0321 09:08:01.535494 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:08:01 crc kubenswrapper[4696]: E0321 09:08:01.535773 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:08:01 crc kubenswrapper[4696]: I0321 09:08:01.799866 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568068-6f9cd" event={"ID":"1bae18e6-4af5-467d-a736-78b4d04a0fa3","Type":"ContainerStarted","Data":"39eb5eed4b2c5dd524e2e11ac75b5323c7f577c2cf04ee1510829a02da30cdb9"} Mar 21 09:08:02 crc kubenswrapper[4696]: I0321 09:08:02.810561 4696 generic.go:334] "Generic (PLEG): container finished" podID="1bae18e6-4af5-467d-a736-78b4d04a0fa3" containerID="baef8d13fe1ecc93a2f2c42df2804e723f27215d40de1d80fd02f6c1da4adc42" exitCode=0 Mar 21 09:08:02 crc kubenswrapper[4696]: I0321 09:08:02.810604 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568068-6f9cd" event={"ID":"1bae18e6-4af5-467d-a736-78b4d04a0fa3","Type":"ContainerDied","Data":"baef8d13fe1ecc93a2f2c42df2804e723f27215d40de1d80fd02f6c1da4adc42"} Mar 21 09:08:04 crc kubenswrapper[4696]: I0321 09:08:04.207558 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568068-6f9cd" Mar 21 09:08:04 crc kubenswrapper[4696]: I0321 09:08:04.274361 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x92fs\" (UniqueName: \"kubernetes.io/projected/1bae18e6-4af5-467d-a736-78b4d04a0fa3-kube-api-access-x92fs\") pod \"1bae18e6-4af5-467d-a736-78b4d04a0fa3\" (UID: \"1bae18e6-4af5-467d-a736-78b4d04a0fa3\") " Mar 21 09:08:04 crc kubenswrapper[4696]: I0321 09:08:04.280133 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bae18e6-4af5-467d-a736-78b4d04a0fa3-kube-api-access-x92fs" (OuterVolumeSpecName: "kube-api-access-x92fs") pod "1bae18e6-4af5-467d-a736-78b4d04a0fa3" (UID: "1bae18e6-4af5-467d-a736-78b4d04a0fa3"). InnerVolumeSpecName "kube-api-access-x92fs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:08:04 crc kubenswrapper[4696]: I0321 09:08:04.377893 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x92fs\" (UniqueName: \"kubernetes.io/projected/1bae18e6-4af5-467d-a736-78b4d04a0fa3-kube-api-access-x92fs\") on node \"crc\" DevicePath \"\"" Mar 21 09:08:04 crc kubenswrapper[4696]: I0321 09:08:04.847230 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568068-6f9cd" event={"ID":"1bae18e6-4af5-467d-a736-78b4d04a0fa3","Type":"ContainerDied","Data":"39eb5eed4b2c5dd524e2e11ac75b5323c7f577c2cf04ee1510829a02da30cdb9"} Mar 21 09:08:04 crc kubenswrapper[4696]: I0321 09:08:04.847286 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39eb5eed4b2c5dd524e2e11ac75b5323c7f577c2cf04ee1510829a02da30cdb9" Mar 21 09:08:04 crc kubenswrapper[4696]: I0321 09:08:04.847287 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568068-6f9cd" Mar 21 09:08:05 crc kubenswrapper[4696]: I0321 09:08:05.277970 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568062-twbpv"] Mar 21 09:08:05 crc kubenswrapper[4696]: I0321 09:08:05.286693 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568062-twbpv"] Mar 21 09:08:06 crc kubenswrapper[4696]: I0321 09:08:06.544896 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1d1e242-e372-472f-a67c-709b9d94dfe4" path="/var/lib/kubelet/pods/d1d1e242-e372-472f-a67c-709b9d94dfe4/volumes" Mar 21 09:08:15 crc kubenswrapper[4696]: I0321 09:08:15.534320 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:08:15 crc kubenswrapper[4696]: E0321 09:08:15.535234 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:08:26 crc kubenswrapper[4696]: I0321 09:08:26.546927 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:08:26 crc kubenswrapper[4696]: E0321 09:08:26.547630 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:08:40 crc kubenswrapper[4696]: I0321 09:08:40.535631 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:08:40 crc kubenswrapper[4696]: E0321 09:08:40.536280 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:08:55 crc kubenswrapper[4696]: I0321 09:08:55.535348 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:08:55 crc kubenswrapper[4696]: E0321 09:08:55.536245 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:09:01 crc kubenswrapper[4696]: I0321 09:09:01.836658 4696 scope.go:117] "RemoveContainer" containerID="630731d5ea50a7b392f5de8cf9f8fca62ee44b1810f5a04d5a8f47cef8109dbb" Mar 21 09:09:09 crc kubenswrapper[4696]: I0321 09:09:09.534370 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:09:09 crc kubenswrapper[4696]: E0321 09:09:09.535172 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:09:23 crc kubenswrapper[4696]: I0321 09:09:23.534628 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:09:23 crc kubenswrapper[4696]: E0321 09:09:23.535455 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:09:23 crc kubenswrapper[4696]: I0321 09:09:23.555705 4696 generic.go:334] "Generic (PLEG): container finished" podID="14ded188-15ec-4d0c-a462-d1160dfb5b47" containerID="ec720d24aa26ccc83152227d846ea6c91cb778773268fefee8acac4e5fb8230f" exitCode=0 Mar 21 09:09:23 crc kubenswrapper[4696]: I0321 09:09:23.555748 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" event={"ID":"14ded188-15ec-4d0c-a462-d1160dfb5b47","Type":"ContainerDied","Data":"ec720d24aa26ccc83152227d846ea6c91cb778773268fefee8acac4e5fb8230f"} Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.044237 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.102053 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-migration-ssh-key-0\") pod \"14ded188-15ec-4d0c-a462-d1160dfb5b47\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.102144 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-0\") pod \"14ded188-15ec-4d0c-a462-d1160dfb5b47\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.102191 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-inventory\") pod \"14ded188-15ec-4d0c-a462-d1160dfb5b47\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.102261 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-combined-ca-bundle\") pod \"14ded188-15ec-4d0c-a462-d1160dfb5b47\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.102301 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-3\") pod \"14ded188-15ec-4d0c-a462-d1160dfb5b47\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.102340 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-1\") pod \"14ded188-15ec-4d0c-a462-d1160dfb5b47\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.102363 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-migration-ssh-key-1\") pod \"14ded188-15ec-4d0c-a462-d1160dfb5b47\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.102396 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-extra-config-0\") pod \"14ded188-15ec-4d0c-a462-d1160dfb5b47\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.102449 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-ssh-key-openstack-edpm-ipam\") pod \"14ded188-15ec-4d0c-a462-d1160dfb5b47\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.102579 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbpvt\" (UniqueName: \"kubernetes.io/projected/14ded188-15ec-4d0c-a462-d1160dfb5b47-kube-api-access-zbpvt\") pod \"14ded188-15ec-4d0c-a462-d1160dfb5b47\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.102681 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-2\") pod \"14ded188-15ec-4d0c-a462-d1160dfb5b47\" (UID: \"14ded188-15ec-4d0c-a462-d1160dfb5b47\") " Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.108690 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "14ded188-15ec-4d0c-a462-d1160dfb5b47" (UID: "14ded188-15ec-4d0c-a462-d1160dfb5b47"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.112681 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14ded188-15ec-4d0c-a462-d1160dfb5b47-kube-api-access-zbpvt" (OuterVolumeSpecName: "kube-api-access-zbpvt") pod "14ded188-15ec-4d0c-a462-d1160dfb5b47" (UID: "14ded188-15ec-4d0c-a462-d1160dfb5b47"). InnerVolumeSpecName "kube-api-access-zbpvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.144309 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "14ded188-15ec-4d0c-a462-d1160dfb5b47" (UID: "14ded188-15ec-4d0c-a462-d1160dfb5b47"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.145094 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "14ded188-15ec-4d0c-a462-d1160dfb5b47" (UID: "14ded188-15ec-4d0c-a462-d1160dfb5b47"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.149980 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "14ded188-15ec-4d0c-a462-d1160dfb5b47" (UID: "14ded188-15ec-4d0c-a462-d1160dfb5b47"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.155802 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "14ded188-15ec-4d0c-a462-d1160dfb5b47" (UID: "14ded188-15ec-4d0c-a462-d1160dfb5b47"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.159958 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "14ded188-15ec-4d0c-a462-d1160dfb5b47" (UID: "14ded188-15ec-4d0c-a462-d1160dfb5b47"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.162031 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "14ded188-15ec-4d0c-a462-d1160dfb5b47" (UID: "14ded188-15ec-4d0c-a462-d1160dfb5b47"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.170125 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "14ded188-15ec-4d0c-a462-d1160dfb5b47" (UID: "14ded188-15ec-4d0c-a462-d1160dfb5b47"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.172891 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-inventory" (OuterVolumeSpecName: "inventory") pod "14ded188-15ec-4d0c-a462-d1160dfb5b47" (UID: "14ded188-15ec-4d0c-a462-d1160dfb5b47"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.180377 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "14ded188-15ec-4d0c-a462-d1160dfb5b47" (UID: "14ded188-15ec-4d0c-a462-d1160dfb5b47"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.205586 4696 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.205620 4696 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.205631 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.205640 4696 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.205649 4696 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.205659 4696 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.205668 4696 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.205678 4696 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.205686 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.205694 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbpvt\" (UniqueName: \"kubernetes.io/projected/14ded188-15ec-4d0c-a462-d1160dfb5b47-kube-api-access-zbpvt\") on node \"crc\" DevicePath \"\"" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.205703 4696 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/14ded188-15ec-4d0c-a462-d1160dfb5b47-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.579637 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" event={"ID":"14ded188-15ec-4d0c-a462-d1160dfb5b47","Type":"ContainerDied","Data":"e0c69d94c374ed50a79270341c43e3ae98d4cde4c67d43bb9af12641cb3ecefb"} Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.579982 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0c69d94c374ed50a79270341c43e3ae98d4cde4c67d43bb9af12641cb3ecefb" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.579714 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dpstl" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.672864 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8"] Mar 21 09:09:25 crc kubenswrapper[4696]: E0321 09:09:25.673350 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ded188-15ec-4d0c-a462-d1160dfb5b47" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.673366 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ded188-15ec-4d0c-a462-d1160dfb5b47" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 21 09:09:25 crc kubenswrapper[4696]: E0321 09:09:25.673404 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bae18e6-4af5-467d-a736-78b4d04a0fa3" containerName="oc" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.673413 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bae18e6-4af5-467d-a736-78b4d04a0fa3" containerName="oc" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.673617 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="14ded188-15ec-4d0c-a462-d1160dfb5b47" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.673631 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bae18e6-4af5-467d-a736-78b4d04a0fa3" containerName="oc" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.674396 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.677347 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.677844 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.678040 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bqp9r" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.678192 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.678330 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.698754 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8"] Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.721639 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.721690 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.721944 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.722041 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.722238 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.722266 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb7hg\" (UniqueName: \"kubernetes.io/projected/2b3f6677-db77-456b-aff2-9deee1fe26f9-kube-api-access-bb7hg\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.722412 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.824771 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.824860 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.824899 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.824916 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb7hg\" (UniqueName: \"kubernetes.io/projected/2b3f6677-db77-456b-aff2-9deee1fe26f9-kube-api-access-bb7hg\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.824962 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.825789 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.826206 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.829191 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.829230 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.829237 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.829562 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.830179 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.836013 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:25 crc kubenswrapper[4696]: I0321 09:09:25.842203 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb7hg\" (UniqueName: \"kubernetes.io/projected/2b3f6677-db77-456b-aff2-9deee1fe26f9-kube-api-access-bb7hg\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:26 crc kubenswrapper[4696]: I0321 09:09:26.035067 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:09:26 crc kubenswrapper[4696]: I0321 09:09:26.557462 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8"] Mar 21 09:09:26 crc kubenswrapper[4696]: W0321 09:09:26.562181 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b3f6677_db77_456b_aff2_9deee1fe26f9.slice/crio-4953ffae4b01d83c805b24d4e75b255748d0ae2d471a6ee9b8cdad7ccbb4f7bd WatchSource:0}: Error finding container 4953ffae4b01d83c805b24d4e75b255748d0ae2d471a6ee9b8cdad7ccbb4f7bd: Status 404 returned error can't find the container with id 4953ffae4b01d83c805b24d4e75b255748d0ae2d471a6ee9b8cdad7ccbb4f7bd Mar 21 09:09:26 crc kubenswrapper[4696]: I0321 09:09:26.589064 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" event={"ID":"2b3f6677-db77-456b-aff2-9deee1fe26f9","Type":"ContainerStarted","Data":"4953ffae4b01d83c805b24d4e75b255748d0ae2d471a6ee9b8cdad7ccbb4f7bd"} Mar 21 09:09:27 crc kubenswrapper[4696]: I0321 09:09:27.602600 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" event={"ID":"2b3f6677-db77-456b-aff2-9deee1fe26f9","Type":"ContainerStarted","Data":"59cb507cbb8551df26c458ea56b342d5faef0e8c290c2be4b34d61ff36362aaf"} Mar 21 09:09:27 crc kubenswrapper[4696]: I0321 09:09:27.627857 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" podStartSLOduration=2.223444499 podStartE2EDuration="2.627838484s" podCreationTimestamp="2026-03-21 09:09:25 +0000 UTC" firstStartedPulling="2026-03-21 09:09:26.565719498 +0000 UTC m=+2500.686600211" lastFinishedPulling="2026-03-21 09:09:26.970113483 +0000 UTC m=+2501.090994196" observedRunningTime="2026-03-21 09:09:27.627409111 +0000 UTC m=+2501.748289864" watchObservedRunningTime="2026-03-21 09:09:27.627838484 +0000 UTC m=+2501.748719197" Mar 21 09:09:38 crc kubenswrapper[4696]: I0321 09:09:38.534794 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:09:38 crc kubenswrapper[4696]: E0321 09:09:38.535567 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:09:46 crc kubenswrapper[4696]: I0321 09:09:46.408568 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cfj6k"] Mar 21 09:09:46 crc kubenswrapper[4696]: I0321 09:09:46.411598 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cfj6k" Mar 21 09:09:46 crc kubenswrapper[4696]: I0321 09:09:46.423360 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cfj6k"] Mar 21 09:09:46 crc kubenswrapper[4696]: I0321 09:09:46.508119 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03e73bed-7124-4873-b10f-ee9b84fd1dca-catalog-content\") pod \"community-operators-cfj6k\" (UID: \"03e73bed-7124-4873-b10f-ee9b84fd1dca\") " pod="openshift-marketplace/community-operators-cfj6k" Mar 21 09:09:46 crc kubenswrapper[4696]: I0321 09:09:46.508357 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbm86\" (UniqueName: \"kubernetes.io/projected/03e73bed-7124-4873-b10f-ee9b84fd1dca-kube-api-access-tbm86\") pod \"community-operators-cfj6k\" (UID: \"03e73bed-7124-4873-b10f-ee9b84fd1dca\") " pod="openshift-marketplace/community-operators-cfj6k" Mar 21 09:09:46 crc kubenswrapper[4696]: I0321 09:09:46.508427 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03e73bed-7124-4873-b10f-ee9b84fd1dca-utilities\") pod \"community-operators-cfj6k\" (UID: \"03e73bed-7124-4873-b10f-ee9b84fd1dca\") " pod="openshift-marketplace/community-operators-cfj6k" Mar 21 09:09:46 crc kubenswrapper[4696]: I0321 09:09:46.610396 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbm86\" (UniqueName: \"kubernetes.io/projected/03e73bed-7124-4873-b10f-ee9b84fd1dca-kube-api-access-tbm86\") pod \"community-operators-cfj6k\" (UID: \"03e73bed-7124-4873-b10f-ee9b84fd1dca\") " pod="openshift-marketplace/community-operators-cfj6k" Mar 21 09:09:46 crc kubenswrapper[4696]: I0321 09:09:46.610729 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03e73bed-7124-4873-b10f-ee9b84fd1dca-utilities\") pod \"community-operators-cfj6k\" (UID: \"03e73bed-7124-4873-b10f-ee9b84fd1dca\") " pod="openshift-marketplace/community-operators-cfj6k" Mar 21 09:09:46 crc kubenswrapper[4696]: I0321 09:09:46.610882 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03e73bed-7124-4873-b10f-ee9b84fd1dca-catalog-content\") pod \"community-operators-cfj6k\" (UID: \"03e73bed-7124-4873-b10f-ee9b84fd1dca\") " pod="openshift-marketplace/community-operators-cfj6k" Mar 21 09:09:46 crc kubenswrapper[4696]: I0321 09:09:46.611530 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03e73bed-7124-4873-b10f-ee9b84fd1dca-utilities\") pod \"community-operators-cfj6k\" (UID: \"03e73bed-7124-4873-b10f-ee9b84fd1dca\") " pod="openshift-marketplace/community-operators-cfj6k" Mar 21 09:09:46 crc kubenswrapper[4696]: I0321 09:09:46.611583 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03e73bed-7124-4873-b10f-ee9b84fd1dca-catalog-content\") pod \"community-operators-cfj6k\" (UID: \"03e73bed-7124-4873-b10f-ee9b84fd1dca\") " pod="openshift-marketplace/community-operators-cfj6k" Mar 21 09:09:46 crc kubenswrapper[4696]: I0321 09:09:46.642250 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbm86\" (UniqueName: \"kubernetes.io/projected/03e73bed-7124-4873-b10f-ee9b84fd1dca-kube-api-access-tbm86\") pod \"community-operators-cfj6k\" (UID: \"03e73bed-7124-4873-b10f-ee9b84fd1dca\") " pod="openshift-marketplace/community-operators-cfj6k" Mar 21 09:09:46 crc kubenswrapper[4696]: I0321 09:09:46.744026 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cfj6k" Mar 21 09:09:47 crc kubenswrapper[4696]: W0321 09:09:47.331474 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03e73bed_7124_4873_b10f_ee9b84fd1dca.slice/crio-b4eae3f4c5096b3bc44bc0d349ab3b3258a14d0eac79abfed1aca72cb8c401da WatchSource:0}: Error finding container b4eae3f4c5096b3bc44bc0d349ab3b3258a14d0eac79abfed1aca72cb8c401da: Status 404 returned error can't find the container with id b4eae3f4c5096b3bc44bc0d349ab3b3258a14d0eac79abfed1aca72cb8c401da Mar 21 09:09:47 crc kubenswrapper[4696]: I0321 09:09:47.335887 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cfj6k"] Mar 21 09:09:47 crc kubenswrapper[4696]: I0321 09:09:47.813266 4696 generic.go:334] "Generic (PLEG): container finished" podID="03e73bed-7124-4873-b10f-ee9b84fd1dca" containerID="1c17bec1540e9dc9a66aedb9f2ed607c9da51200e75dcbc7220e83f7f64cf17e" exitCode=0 Mar 21 09:09:47 crc kubenswrapper[4696]: I0321 09:09:47.813339 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cfj6k" event={"ID":"03e73bed-7124-4873-b10f-ee9b84fd1dca","Type":"ContainerDied","Data":"1c17bec1540e9dc9a66aedb9f2ed607c9da51200e75dcbc7220e83f7f64cf17e"} Mar 21 09:09:47 crc kubenswrapper[4696]: I0321 09:09:47.813650 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cfj6k" event={"ID":"03e73bed-7124-4873-b10f-ee9b84fd1dca","Type":"ContainerStarted","Data":"b4eae3f4c5096b3bc44bc0d349ab3b3258a14d0eac79abfed1aca72cb8c401da"} Mar 21 09:09:48 crc kubenswrapper[4696]: I0321 09:09:48.825770 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cfj6k" event={"ID":"03e73bed-7124-4873-b10f-ee9b84fd1dca","Type":"ContainerStarted","Data":"78eae54209706990cdf707b4c7b9a44ef714332a99ddf9d288fbd4cd9b85df0b"} Mar 21 09:09:50 crc kubenswrapper[4696]: I0321 09:09:50.864704 4696 generic.go:334] "Generic (PLEG): container finished" podID="03e73bed-7124-4873-b10f-ee9b84fd1dca" containerID="78eae54209706990cdf707b4c7b9a44ef714332a99ddf9d288fbd4cd9b85df0b" exitCode=0 Mar 21 09:09:50 crc kubenswrapper[4696]: I0321 09:09:50.864798 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cfj6k" event={"ID":"03e73bed-7124-4873-b10f-ee9b84fd1dca","Type":"ContainerDied","Data":"78eae54209706990cdf707b4c7b9a44ef714332a99ddf9d288fbd4cd9b85df0b"} Mar 21 09:09:51 crc kubenswrapper[4696]: I0321 09:09:51.535031 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:09:51 crc kubenswrapper[4696]: E0321 09:09:51.535634 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:09:51 crc kubenswrapper[4696]: I0321 09:09:51.876006 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cfj6k" event={"ID":"03e73bed-7124-4873-b10f-ee9b84fd1dca","Type":"ContainerStarted","Data":"b415cfefc7b9484d776c96e5a9842d26cb8d94c4a27910cb1b1e92935775350f"} Mar 21 09:09:51 crc kubenswrapper[4696]: I0321 09:09:51.896431 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cfj6k" podStartSLOduration=2.448026089 podStartE2EDuration="5.896413461s" podCreationTimestamp="2026-03-21 09:09:46 +0000 UTC" firstStartedPulling="2026-03-21 09:09:47.816106538 +0000 UTC m=+2521.936987261" lastFinishedPulling="2026-03-21 09:09:51.26449391 +0000 UTC m=+2525.385374633" observedRunningTime="2026-03-21 09:09:51.891142035 +0000 UTC m=+2526.012022748" watchObservedRunningTime="2026-03-21 09:09:51.896413461 +0000 UTC m=+2526.017294174" Mar 21 09:09:56 crc kubenswrapper[4696]: I0321 09:09:56.744658 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cfj6k" Mar 21 09:09:56 crc kubenswrapper[4696]: I0321 09:09:56.745284 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cfj6k" Mar 21 09:09:56 crc kubenswrapper[4696]: I0321 09:09:56.795375 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cfj6k" Mar 21 09:09:56 crc kubenswrapper[4696]: I0321 09:09:56.980143 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cfj6k" Mar 21 09:10:00 crc kubenswrapper[4696]: I0321 09:10:00.143360 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568070-sp48z"] Mar 21 09:10:00 crc kubenswrapper[4696]: I0321 09:10:00.145344 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568070-sp48z" Mar 21 09:10:00 crc kubenswrapper[4696]: I0321 09:10:00.147285 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:10:00 crc kubenswrapper[4696]: I0321 09:10:00.147861 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:10:00 crc kubenswrapper[4696]: I0321 09:10:00.148641 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:10:00 crc kubenswrapper[4696]: I0321 09:10:00.152683 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568070-sp48z"] Mar 21 09:10:00 crc kubenswrapper[4696]: I0321 09:10:00.288678 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2frt\" (UniqueName: \"kubernetes.io/projected/303575eb-eb75-4bf4-8501-5e67a2c72470-kube-api-access-p2frt\") pod \"auto-csr-approver-29568070-sp48z\" (UID: \"303575eb-eb75-4bf4-8501-5e67a2c72470\") " pod="openshift-infra/auto-csr-approver-29568070-sp48z" Mar 21 09:10:00 crc kubenswrapper[4696]: I0321 09:10:00.391511 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2frt\" (UniqueName: \"kubernetes.io/projected/303575eb-eb75-4bf4-8501-5e67a2c72470-kube-api-access-p2frt\") pod \"auto-csr-approver-29568070-sp48z\" (UID: \"303575eb-eb75-4bf4-8501-5e67a2c72470\") " pod="openshift-infra/auto-csr-approver-29568070-sp48z" Mar 21 09:10:00 crc kubenswrapper[4696]: I0321 09:10:00.401383 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cfj6k"] Mar 21 09:10:00 crc kubenswrapper[4696]: I0321 09:10:00.401742 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cfj6k" podUID="03e73bed-7124-4873-b10f-ee9b84fd1dca" containerName="registry-server" containerID="cri-o://b415cfefc7b9484d776c96e5a9842d26cb8d94c4a27910cb1b1e92935775350f" gracePeriod=2 Mar 21 09:10:00 crc kubenswrapper[4696]: I0321 09:10:00.415012 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2frt\" (UniqueName: \"kubernetes.io/projected/303575eb-eb75-4bf4-8501-5e67a2c72470-kube-api-access-p2frt\") pod \"auto-csr-approver-29568070-sp48z\" (UID: \"303575eb-eb75-4bf4-8501-5e67a2c72470\") " pod="openshift-infra/auto-csr-approver-29568070-sp48z" Mar 21 09:10:00 crc kubenswrapper[4696]: I0321 09:10:00.468628 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568070-sp48z" Mar 21 09:10:00 crc kubenswrapper[4696]: I0321 09:10:00.939394 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cfj6k" Mar 21 09:10:00 crc kubenswrapper[4696]: I0321 09:10:00.967362 4696 generic.go:334] "Generic (PLEG): container finished" podID="03e73bed-7124-4873-b10f-ee9b84fd1dca" containerID="b415cfefc7b9484d776c96e5a9842d26cb8d94c4a27910cb1b1e92935775350f" exitCode=0 Mar 21 09:10:00 crc kubenswrapper[4696]: I0321 09:10:00.967693 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cfj6k" Mar 21 09:10:00 crc kubenswrapper[4696]: I0321 09:10:00.967706 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cfj6k" event={"ID":"03e73bed-7124-4873-b10f-ee9b84fd1dca","Type":"ContainerDied","Data":"b415cfefc7b9484d776c96e5a9842d26cb8d94c4a27910cb1b1e92935775350f"} Mar 21 09:10:00 crc kubenswrapper[4696]: I0321 09:10:00.968504 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cfj6k" event={"ID":"03e73bed-7124-4873-b10f-ee9b84fd1dca","Type":"ContainerDied","Data":"b4eae3f4c5096b3bc44bc0d349ab3b3258a14d0eac79abfed1aca72cb8c401da"} Mar 21 09:10:00 crc kubenswrapper[4696]: I0321 09:10:00.968544 4696 scope.go:117] "RemoveContainer" containerID="b415cfefc7b9484d776c96e5a9842d26cb8d94c4a27910cb1b1e92935775350f" Mar 21 09:10:00 crc kubenswrapper[4696]: I0321 09:10:00.995889 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568070-sp48z"] Mar 21 09:10:00 crc kubenswrapper[4696]: I0321 09:10:00.997133 4696 scope.go:117] "RemoveContainer" containerID="78eae54209706990cdf707b4c7b9a44ef714332a99ddf9d288fbd4cd9b85df0b" Mar 21 09:10:01 crc kubenswrapper[4696]: W0321 09:10:01.005361 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod303575eb_eb75_4bf4_8501_5e67a2c72470.slice/crio-451fa565817180b04d797dd07ed2507a8968a8da752aeffb346a9d4eaeca4634 WatchSource:0}: Error finding container 451fa565817180b04d797dd07ed2507a8968a8da752aeffb346a9d4eaeca4634: Status 404 returned error can't find the container with id 451fa565817180b04d797dd07ed2507a8968a8da752aeffb346a9d4eaeca4634 Mar 21 09:10:01 crc kubenswrapper[4696]: I0321 09:10:01.012562 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03e73bed-7124-4873-b10f-ee9b84fd1dca-catalog-content\") pod \"03e73bed-7124-4873-b10f-ee9b84fd1dca\" (UID: \"03e73bed-7124-4873-b10f-ee9b84fd1dca\") " Mar 21 09:10:01 crc kubenswrapper[4696]: I0321 09:10:01.012859 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbm86\" (UniqueName: \"kubernetes.io/projected/03e73bed-7124-4873-b10f-ee9b84fd1dca-kube-api-access-tbm86\") pod \"03e73bed-7124-4873-b10f-ee9b84fd1dca\" (UID: \"03e73bed-7124-4873-b10f-ee9b84fd1dca\") " Mar 21 09:10:01 crc kubenswrapper[4696]: I0321 09:10:01.012929 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03e73bed-7124-4873-b10f-ee9b84fd1dca-utilities\") pod \"03e73bed-7124-4873-b10f-ee9b84fd1dca\" (UID: \"03e73bed-7124-4873-b10f-ee9b84fd1dca\") " Mar 21 09:10:01 crc kubenswrapper[4696]: I0321 09:10:01.015506 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03e73bed-7124-4873-b10f-ee9b84fd1dca-utilities" (OuterVolumeSpecName: "utilities") pod "03e73bed-7124-4873-b10f-ee9b84fd1dca" (UID: "03e73bed-7124-4873-b10f-ee9b84fd1dca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:10:01 crc kubenswrapper[4696]: I0321 09:10:01.019335 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03e73bed-7124-4873-b10f-ee9b84fd1dca-kube-api-access-tbm86" (OuterVolumeSpecName: "kube-api-access-tbm86") pod "03e73bed-7124-4873-b10f-ee9b84fd1dca" (UID: "03e73bed-7124-4873-b10f-ee9b84fd1dca"). InnerVolumeSpecName "kube-api-access-tbm86". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:10:01 crc kubenswrapper[4696]: I0321 09:10:01.060050 4696 scope.go:117] "RemoveContainer" containerID="1c17bec1540e9dc9a66aedb9f2ed607c9da51200e75dcbc7220e83f7f64cf17e" Mar 21 09:10:01 crc kubenswrapper[4696]: I0321 09:10:01.068756 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03e73bed-7124-4873-b10f-ee9b84fd1dca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03e73bed-7124-4873-b10f-ee9b84fd1dca" (UID: "03e73bed-7124-4873-b10f-ee9b84fd1dca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:10:01 crc kubenswrapper[4696]: I0321 09:10:01.079271 4696 scope.go:117] "RemoveContainer" containerID="b415cfefc7b9484d776c96e5a9842d26cb8d94c4a27910cb1b1e92935775350f" Mar 21 09:10:01 crc kubenswrapper[4696]: E0321 09:10:01.079700 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b415cfefc7b9484d776c96e5a9842d26cb8d94c4a27910cb1b1e92935775350f\": container with ID starting with b415cfefc7b9484d776c96e5a9842d26cb8d94c4a27910cb1b1e92935775350f not found: ID does not exist" containerID="b415cfefc7b9484d776c96e5a9842d26cb8d94c4a27910cb1b1e92935775350f" Mar 21 09:10:01 crc kubenswrapper[4696]: I0321 09:10:01.079768 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b415cfefc7b9484d776c96e5a9842d26cb8d94c4a27910cb1b1e92935775350f"} err="failed to get container status \"b415cfefc7b9484d776c96e5a9842d26cb8d94c4a27910cb1b1e92935775350f\": rpc error: code = NotFound desc = could not find container \"b415cfefc7b9484d776c96e5a9842d26cb8d94c4a27910cb1b1e92935775350f\": container with ID starting with b415cfefc7b9484d776c96e5a9842d26cb8d94c4a27910cb1b1e92935775350f not found: ID does not exist" Mar 21 09:10:01 crc kubenswrapper[4696]: I0321 09:10:01.079798 4696 scope.go:117] "RemoveContainer" containerID="78eae54209706990cdf707b4c7b9a44ef714332a99ddf9d288fbd4cd9b85df0b" Mar 21 09:10:01 crc kubenswrapper[4696]: E0321 09:10:01.080287 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78eae54209706990cdf707b4c7b9a44ef714332a99ddf9d288fbd4cd9b85df0b\": container with ID starting with 78eae54209706990cdf707b4c7b9a44ef714332a99ddf9d288fbd4cd9b85df0b not found: ID does not exist" containerID="78eae54209706990cdf707b4c7b9a44ef714332a99ddf9d288fbd4cd9b85df0b" Mar 21 09:10:01 crc kubenswrapper[4696]: I0321 09:10:01.080316 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78eae54209706990cdf707b4c7b9a44ef714332a99ddf9d288fbd4cd9b85df0b"} err="failed to get container status \"78eae54209706990cdf707b4c7b9a44ef714332a99ddf9d288fbd4cd9b85df0b\": rpc error: code = NotFound desc = could not find container \"78eae54209706990cdf707b4c7b9a44ef714332a99ddf9d288fbd4cd9b85df0b\": container with ID starting with 78eae54209706990cdf707b4c7b9a44ef714332a99ddf9d288fbd4cd9b85df0b not found: ID does not exist" Mar 21 09:10:01 crc kubenswrapper[4696]: I0321 09:10:01.080335 4696 scope.go:117] "RemoveContainer" containerID="1c17bec1540e9dc9a66aedb9f2ed607c9da51200e75dcbc7220e83f7f64cf17e" Mar 21 09:10:01 crc kubenswrapper[4696]: E0321 09:10:01.080576 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c17bec1540e9dc9a66aedb9f2ed607c9da51200e75dcbc7220e83f7f64cf17e\": container with ID starting with 1c17bec1540e9dc9a66aedb9f2ed607c9da51200e75dcbc7220e83f7f64cf17e not found: ID does not exist" containerID="1c17bec1540e9dc9a66aedb9f2ed607c9da51200e75dcbc7220e83f7f64cf17e" Mar 21 09:10:01 crc kubenswrapper[4696]: I0321 09:10:01.080605 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c17bec1540e9dc9a66aedb9f2ed607c9da51200e75dcbc7220e83f7f64cf17e"} err="failed to get container status \"1c17bec1540e9dc9a66aedb9f2ed607c9da51200e75dcbc7220e83f7f64cf17e\": rpc error: code = NotFound desc = could not find container \"1c17bec1540e9dc9a66aedb9f2ed607c9da51200e75dcbc7220e83f7f64cf17e\": container with ID starting with 1c17bec1540e9dc9a66aedb9f2ed607c9da51200e75dcbc7220e83f7f64cf17e not found: ID does not exist" Mar 21 09:10:01 crc kubenswrapper[4696]: I0321 09:10:01.115851 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbm86\" (UniqueName: \"kubernetes.io/projected/03e73bed-7124-4873-b10f-ee9b84fd1dca-kube-api-access-tbm86\") on node \"crc\" DevicePath \"\"" Mar 21 09:10:01 crc kubenswrapper[4696]: I0321 09:10:01.115878 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03e73bed-7124-4873-b10f-ee9b84fd1dca-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 09:10:01 crc kubenswrapper[4696]: I0321 09:10:01.115888 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03e73bed-7124-4873-b10f-ee9b84fd1dca-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 09:10:01 crc kubenswrapper[4696]: I0321 09:10:01.313221 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cfj6k"] Mar 21 09:10:01 crc kubenswrapper[4696]: I0321 09:10:01.322786 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cfj6k"] Mar 21 09:10:01 crc kubenswrapper[4696]: I0321 09:10:01.976342 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568070-sp48z" event={"ID":"303575eb-eb75-4bf4-8501-5e67a2c72470","Type":"ContainerStarted","Data":"451fa565817180b04d797dd07ed2507a8968a8da752aeffb346a9d4eaeca4634"} Mar 21 09:10:02 crc kubenswrapper[4696]: I0321 09:10:02.535193 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:10:02 crc kubenswrapper[4696]: E0321 09:10:02.535681 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:10:02 crc kubenswrapper[4696]: I0321 09:10:02.547342 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03e73bed-7124-4873-b10f-ee9b84fd1dca" path="/var/lib/kubelet/pods/03e73bed-7124-4873-b10f-ee9b84fd1dca/volumes" Mar 21 09:10:02 crc kubenswrapper[4696]: I0321 09:10:02.986769 4696 generic.go:334] "Generic (PLEG): container finished" podID="303575eb-eb75-4bf4-8501-5e67a2c72470" containerID="1dd24cefb0132751f500df4ae16247679726ae6ff291d82371f5ed45d48b3d7f" exitCode=0 Mar 21 09:10:02 crc kubenswrapper[4696]: I0321 09:10:02.986857 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568070-sp48z" event={"ID":"303575eb-eb75-4bf4-8501-5e67a2c72470","Type":"ContainerDied","Data":"1dd24cefb0132751f500df4ae16247679726ae6ff291d82371f5ed45d48b3d7f"} Mar 21 09:10:04 crc kubenswrapper[4696]: I0321 09:10:04.351476 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568070-sp48z" Mar 21 09:10:04 crc kubenswrapper[4696]: I0321 09:10:04.481897 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2frt\" (UniqueName: \"kubernetes.io/projected/303575eb-eb75-4bf4-8501-5e67a2c72470-kube-api-access-p2frt\") pod \"303575eb-eb75-4bf4-8501-5e67a2c72470\" (UID: \"303575eb-eb75-4bf4-8501-5e67a2c72470\") " Mar 21 09:10:04 crc kubenswrapper[4696]: I0321 09:10:04.490040 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/303575eb-eb75-4bf4-8501-5e67a2c72470-kube-api-access-p2frt" (OuterVolumeSpecName: "kube-api-access-p2frt") pod "303575eb-eb75-4bf4-8501-5e67a2c72470" (UID: "303575eb-eb75-4bf4-8501-5e67a2c72470"). InnerVolumeSpecName "kube-api-access-p2frt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:10:04 crc kubenswrapper[4696]: I0321 09:10:04.584073 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2frt\" (UniqueName: \"kubernetes.io/projected/303575eb-eb75-4bf4-8501-5e67a2c72470-kube-api-access-p2frt\") on node \"crc\" DevicePath \"\"" Mar 21 09:10:05 crc kubenswrapper[4696]: I0321 09:10:05.005772 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568070-sp48z" event={"ID":"303575eb-eb75-4bf4-8501-5e67a2c72470","Type":"ContainerDied","Data":"451fa565817180b04d797dd07ed2507a8968a8da752aeffb346a9d4eaeca4634"} Mar 21 09:10:05 crc kubenswrapper[4696]: I0321 09:10:05.006033 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="451fa565817180b04d797dd07ed2507a8968a8da752aeffb346a9d4eaeca4634" Mar 21 09:10:05 crc kubenswrapper[4696]: I0321 09:10:05.005876 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568070-sp48z" Mar 21 09:10:05 crc kubenswrapper[4696]: I0321 09:10:05.419167 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568064-lq59p"] Mar 21 09:10:05 crc kubenswrapper[4696]: I0321 09:10:05.428947 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568064-lq59p"] Mar 21 09:10:06 crc kubenswrapper[4696]: I0321 09:10:06.545301 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d52bfab3-d526-433b-86ac-3213f08251d6" path="/var/lib/kubelet/pods/d52bfab3-d526-433b-86ac-3213f08251d6/volumes" Mar 21 09:10:16 crc kubenswrapper[4696]: I0321 09:10:16.544405 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:10:16 crc kubenswrapper[4696]: E0321 09:10:16.545193 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:10:30 crc kubenswrapper[4696]: I0321 09:10:30.534597 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:10:30 crc kubenswrapper[4696]: E0321 09:10:30.535571 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:10:41 crc kubenswrapper[4696]: I0321 09:10:41.534783 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:10:41 crc kubenswrapper[4696]: E0321 09:10:41.535504 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:10:52 crc kubenswrapper[4696]: I0321 09:10:52.534862 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:10:52 crc kubenswrapper[4696]: E0321 09:10:52.535637 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:11:01 crc kubenswrapper[4696]: I0321 09:11:01.940974 4696 scope.go:117] "RemoveContainer" containerID="bf3c73e483decd43c6b32dfde33125cf9a48de112affaa939958925a8bb1c3c3" Mar 21 09:11:04 crc kubenswrapper[4696]: I0321 09:11:04.534724 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:11:04 crc kubenswrapper[4696]: E0321 09:11:04.535279 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:11:19 crc kubenswrapper[4696]: I0321 09:11:19.535159 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:11:19 crc kubenswrapper[4696]: E0321 09:11:19.535993 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:11:34 crc kubenswrapper[4696]: I0321 09:11:34.534687 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:11:34 crc kubenswrapper[4696]: I0321 09:11:34.794666 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerStarted","Data":"a2f3653fbac2953f5711ffc6be0faa87b5c64ec53ea926ab1544f651d7adfcdd"} Mar 21 09:11:45 crc kubenswrapper[4696]: I0321 09:11:45.900169 4696 generic.go:334] "Generic (PLEG): container finished" podID="2b3f6677-db77-456b-aff2-9deee1fe26f9" containerID="59cb507cbb8551df26c458ea56b342d5faef0e8c290c2be4b34d61ff36362aaf" exitCode=0 Mar 21 09:11:45 crc kubenswrapper[4696]: I0321 09:11:45.900255 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" event={"ID":"2b3f6677-db77-456b-aff2-9deee1fe26f9","Type":"ContainerDied","Data":"59cb507cbb8551df26c458ea56b342d5faef0e8c290c2be4b34d61ff36362aaf"} Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.443490 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.563264 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-inventory\") pod \"2b3f6677-db77-456b-aff2-9deee1fe26f9\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.563578 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ssh-key-openstack-edpm-ipam\") pod \"2b3f6677-db77-456b-aff2-9deee1fe26f9\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.563685 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-telemetry-combined-ca-bundle\") pod \"2b3f6677-db77-456b-aff2-9deee1fe26f9\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.563882 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ceilometer-compute-config-data-1\") pod \"2b3f6677-db77-456b-aff2-9deee1fe26f9\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.563992 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ceilometer-compute-config-data-2\") pod \"2b3f6677-db77-456b-aff2-9deee1fe26f9\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.564135 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ceilometer-compute-config-data-0\") pod \"2b3f6677-db77-456b-aff2-9deee1fe26f9\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.564226 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bb7hg\" (UniqueName: \"kubernetes.io/projected/2b3f6677-db77-456b-aff2-9deee1fe26f9-kube-api-access-bb7hg\") pod \"2b3f6677-db77-456b-aff2-9deee1fe26f9\" (UID: \"2b3f6677-db77-456b-aff2-9deee1fe26f9\") " Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.569247 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "2b3f6677-db77-456b-aff2-9deee1fe26f9" (UID: "2b3f6677-db77-456b-aff2-9deee1fe26f9"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.569456 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b3f6677-db77-456b-aff2-9deee1fe26f9-kube-api-access-bb7hg" (OuterVolumeSpecName: "kube-api-access-bb7hg") pod "2b3f6677-db77-456b-aff2-9deee1fe26f9" (UID: "2b3f6677-db77-456b-aff2-9deee1fe26f9"). InnerVolumeSpecName "kube-api-access-bb7hg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.593007 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "2b3f6677-db77-456b-aff2-9deee1fe26f9" (UID: "2b3f6677-db77-456b-aff2-9deee1fe26f9"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.595226 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "2b3f6677-db77-456b-aff2-9deee1fe26f9" (UID: "2b3f6677-db77-456b-aff2-9deee1fe26f9"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.595763 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "2b3f6677-db77-456b-aff2-9deee1fe26f9" (UID: "2b3f6677-db77-456b-aff2-9deee1fe26f9"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.609722 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-inventory" (OuterVolumeSpecName: "inventory") pod "2b3f6677-db77-456b-aff2-9deee1fe26f9" (UID: "2b3f6677-db77-456b-aff2-9deee1fe26f9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.619468 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2b3f6677-db77-456b-aff2-9deee1fe26f9" (UID: "2b3f6677-db77-456b-aff2-9deee1fe26f9"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.666748 4696 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.666782 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bb7hg\" (UniqueName: \"kubernetes.io/projected/2b3f6677-db77-456b-aff2-9deee1fe26f9-kube-api-access-bb7hg\") on node \"crc\" DevicePath \"\"" Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.666795 4696 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-inventory\") on node \"crc\" DevicePath \"\"" Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.666804 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.666830 4696 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.666840 4696 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.666849 4696 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/2b3f6677-db77-456b-aff2-9deee1fe26f9-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.923729 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" event={"ID":"2b3f6677-db77-456b-aff2-9deee1fe26f9","Type":"ContainerDied","Data":"4953ffae4b01d83c805b24d4e75b255748d0ae2d471a6ee9b8cdad7ccbb4f7bd"} Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.923775 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4953ffae4b01d83c805b24d4e75b255748d0ae2d471a6ee9b8cdad7ccbb4f7bd" Mar 21 09:11:47 crc kubenswrapper[4696]: I0321 09:11:47.923800 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8" Mar 21 09:12:00 crc kubenswrapper[4696]: I0321 09:12:00.152196 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568072-nkw88"] Mar 21 09:12:00 crc kubenswrapper[4696]: E0321 09:12:00.153121 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03e73bed-7124-4873-b10f-ee9b84fd1dca" containerName="extract-utilities" Mar 21 09:12:00 crc kubenswrapper[4696]: I0321 09:12:00.153136 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="03e73bed-7124-4873-b10f-ee9b84fd1dca" containerName="extract-utilities" Mar 21 09:12:00 crc kubenswrapper[4696]: E0321 09:12:00.153164 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="303575eb-eb75-4bf4-8501-5e67a2c72470" containerName="oc" Mar 21 09:12:00 crc kubenswrapper[4696]: I0321 09:12:00.153172 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="303575eb-eb75-4bf4-8501-5e67a2c72470" containerName="oc" Mar 21 09:12:00 crc kubenswrapper[4696]: E0321 09:12:00.153202 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03e73bed-7124-4873-b10f-ee9b84fd1dca" containerName="registry-server" Mar 21 09:12:00 crc kubenswrapper[4696]: I0321 09:12:00.153210 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="03e73bed-7124-4873-b10f-ee9b84fd1dca" containerName="registry-server" Mar 21 09:12:00 crc kubenswrapper[4696]: E0321 09:12:00.153221 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03e73bed-7124-4873-b10f-ee9b84fd1dca" containerName="extract-content" Mar 21 09:12:00 crc kubenswrapper[4696]: I0321 09:12:00.153229 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="03e73bed-7124-4873-b10f-ee9b84fd1dca" containerName="extract-content" Mar 21 09:12:00 crc kubenswrapper[4696]: E0321 09:12:00.153240 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b3f6677-db77-456b-aff2-9deee1fe26f9" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 21 09:12:00 crc kubenswrapper[4696]: I0321 09:12:00.153249 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b3f6677-db77-456b-aff2-9deee1fe26f9" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 21 09:12:00 crc kubenswrapper[4696]: I0321 09:12:00.153465 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="03e73bed-7124-4873-b10f-ee9b84fd1dca" containerName="registry-server" Mar 21 09:12:00 crc kubenswrapper[4696]: I0321 09:12:00.153488 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="303575eb-eb75-4bf4-8501-5e67a2c72470" containerName="oc" Mar 21 09:12:00 crc kubenswrapper[4696]: I0321 09:12:00.153514 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b3f6677-db77-456b-aff2-9deee1fe26f9" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 21 09:12:00 crc kubenswrapper[4696]: I0321 09:12:00.154372 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568072-nkw88" Mar 21 09:12:00 crc kubenswrapper[4696]: I0321 09:12:00.157647 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:12:00 crc kubenswrapper[4696]: I0321 09:12:00.157647 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:12:00 crc kubenswrapper[4696]: I0321 09:12:00.158356 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:12:00 crc kubenswrapper[4696]: I0321 09:12:00.166360 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568072-nkw88"] Mar 21 09:12:00 crc kubenswrapper[4696]: I0321 09:12:00.350666 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2w9d\" (UniqueName: \"kubernetes.io/projected/09438c8e-71b8-4bdb-a32e-e07d0d641b0a-kube-api-access-b2w9d\") pod \"auto-csr-approver-29568072-nkw88\" (UID: \"09438c8e-71b8-4bdb-a32e-e07d0d641b0a\") " pod="openshift-infra/auto-csr-approver-29568072-nkw88" Mar 21 09:12:00 crc kubenswrapper[4696]: I0321 09:12:00.453542 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2w9d\" (UniqueName: \"kubernetes.io/projected/09438c8e-71b8-4bdb-a32e-e07d0d641b0a-kube-api-access-b2w9d\") pod \"auto-csr-approver-29568072-nkw88\" (UID: \"09438c8e-71b8-4bdb-a32e-e07d0d641b0a\") " pod="openshift-infra/auto-csr-approver-29568072-nkw88" Mar 21 09:12:00 crc kubenswrapper[4696]: I0321 09:12:00.473701 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2w9d\" (UniqueName: \"kubernetes.io/projected/09438c8e-71b8-4bdb-a32e-e07d0d641b0a-kube-api-access-b2w9d\") pod \"auto-csr-approver-29568072-nkw88\" (UID: \"09438c8e-71b8-4bdb-a32e-e07d0d641b0a\") " pod="openshift-infra/auto-csr-approver-29568072-nkw88" Mar 21 09:12:00 crc kubenswrapper[4696]: I0321 09:12:00.482650 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568072-nkw88" Mar 21 09:12:00 crc kubenswrapper[4696]: I0321 09:12:00.942802 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568072-nkw88"] Mar 21 09:12:01 crc kubenswrapper[4696]: I0321 09:12:01.058592 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568072-nkw88" event={"ID":"09438c8e-71b8-4bdb-a32e-e07d0d641b0a","Type":"ContainerStarted","Data":"02b663c8a0280309fe89cfd7a02bf1cfb2a8072e46a6aa234b05be0b5171b8c4"} Mar 21 09:12:03 crc kubenswrapper[4696]: I0321 09:12:03.080629 4696 generic.go:334] "Generic (PLEG): container finished" podID="09438c8e-71b8-4bdb-a32e-e07d0d641b0a" containerID="94f55f343f50a788ca0600e9a5903a4bdd94d7990d1db08b260ef92786b722ee" exitCode=0 Mar 21 09:12:03 crc kubenswrapper[4696]: I0321 09:12:03.080722 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568072-nkw88" event={"ID":"09438c8e-71b8-4bdb-a32e-e07d0d641b0a","Type":"ContainerDied","Data":"94f55f343f50a788ca0600e9a5903a4bdd94d7990d1db08b260ef92786b722ee"} Mar 21 09:12:04 crc kubenswrapper[4696]: I0321 09:12:04.522059 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568072-nkw88" Mar 21 09:12:04 crc kubenswrapper[4696]: I0321 09:12:04.642079 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2w9d\" (UniqueName: \"kubernetes.io/projected/09438c8e-71b8-4bdb-a32e-e07d0d641b0a-kube-api-access-b2w9d\") pod \"09438c8e-71b8-4bdb-a32e-e07d0d641b0a\" (UID: \"09438c8e-71b8-4bdb-a32e-e07d0d641b0a\") " Mar 21 09:12:04 crc kubenswrapper[4696]: I0321 09:12:04.654026 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09438c8e-71b8-4bdb-a32e-e07d0d641b0a-kube-api-access-b2w9d" (OuterVolumeSpecName: "kube-api-access-b2w9d") pod "09438c8e-71b8-4bdb-a32e-e07d0d641b0a" (UID: "09438c8e-71b8-4bdb-a32e-e07d0d641b0a"). InnerVolumeSpecName "kube-api-access-b2w9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:12:04 crc kubenswrapper[4696]: I0321 09:12:04.744331 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2w9d\" (UniqueName: \"kubernetes.io/projected/09438c8e-71b8-4bdb-a32e-e07d0d641b0a-kube-api-access-b2w9d\") on node \"crc\" DevicePath \"\"" Mar 21 09:12:05 crc kubenswrapper[4696]: I0321 09:12:05.121044 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568072-nkw88" event={"ID":"09438c8e-71b8-4bdb-a32e-e07d0d641b0a","Type":"ContainerDied","Data":"02b663c8a0280309fe89cfd7a02bf1cfb2a8072e46a6aa234b05be0b5171b8c4"} Mar 21 09:12:05 crc kubenswrapper[4696]: I0321 09:12:05.121117 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02b663c8a0280309fe89cfd7a02bf1cfb2a8072e46a6aa234b05be0b5171b8c4" Mar 21 09:12:05 crc kubenswrapper[4696]: I0321 09:12:05.121163 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568072-nkw88" Mar 21 09:12:05 crc kubenswrapper[4696]: I0321 09:12:05.597715 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568066-8qgbv"] Mar 21 09:12:05 crc kubenswrapper[4696]: I0321 09:12:05.608090 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568066-8qgbv"] Mar 21 09:12:06 crc kubenswrapper[4696]: I0321 09:12:06.548582 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5721fef-e6a4-4e01-a722-5436a7ce7eff" path="/var/lib/kubelet/pods/e5721fef-e6a4-4e01-a722-5436a7ce7eff/volumes" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.697831 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Mar 21 09:12:39 crc kubenswrapper[4696]: E0321 09:12:39.698867 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09438c8e-71b8-4bdb-a32e-e07d0d641b0a" containerName="oc" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.698883 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="09438c8e-71b8-4bdb-a32e-e07d0d641b0a" containerName="oc" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.699097 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="09438c8e-71b8-4bdb-a32e-e07d0d641b0a" containerName="oc" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.700034 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.702863 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.702937 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.703577 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.703687 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-cdj78" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.716648 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.832375 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.832706 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.832753 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.832986 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.833063 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx624\" (UniqueName: \"kubernetes.io/projected/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-kube-api-access-kx624\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.833098 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-config-data\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.833135 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.833184 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.833239 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.935252 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.935325 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.935399 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.935425 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx624\" (UniqueName: \"kubernetes.io/projected/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-kube-api-access-kx624\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.935442 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-config-data\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.935460 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.935482 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.935504 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.935594 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.935888 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.936936 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.937277 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.939742 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.940362 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.944789 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.951019 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.952944 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-config-data\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.955910 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx624\" (UniqueName: \"kubernetes.io/projected/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-kube-api-access-kx624\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:39 crc kubenswrapper[4696]: I0321 09:12:39.971566 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " pod="openstack/tempest-tests-tempest" Mar 21 09:12:40 crc kubenswrapper[4696]: I0321 09:12:40.023929 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 21 09:12:40 crc kubenswrapper[4696]: I0321 09:12:40.525161 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Mar 21 09:12:40 crc kubenswrapper[4696]: W0321 09:12:40.529064 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a6c65a7_b20f_4675_a6fc_c4e1988a2a1e.slice/crio-4820203f47462f9c4ceaaa5022751be45b3923440b83ad12d76a1afcdf34a416 WatchSource:0}: Error finding container 4820203f47462f9c4ceaaa5022751be45b3923440b83ad12d76a1afcdf34a416: Status 404 returned error can't find the container with id 4820203f47462f9c4ceaaa5022751be45b3923440b83ad12d76a1afcdf34a416 Mar 21 09:12:41 crc kubenswrapper[4696]: I0321 09:12:41.498972 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e","Type":"ContainerStarted","Data":"4820203f47462f9c4ceaaa5022751be45b3923440b83ad12d76a1afcdf34a416"} Mar 21 09:12:51 crc kubenswrapper[4696]: I0321 09:12:51.520605 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jbt2w"] Mar 21 09:12:51 crc kubenswrapper[4696]: I0321 09:12:51.523418 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbt2w" Mar 21 09:12:51 crc kubenswrapper[4696]: I0321 09:12:51.532975 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jbt2w"] Mar 21 09:12:51 crc kubenswrapper[4696]: I0321 09:12:51.583966 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d6364e4-44f2-42c2-b980-1ad5c1faba7b-catalog-content\") pod \"redhat-operators-jbt2w\" (UID: \"7d6364e4-44f2-42c2-b980-1ad5c1faba7b\") " pod="openshift-marketplace/redhat-operators-jbt2w" Mar 21 09:12:51 crc kubenswrapper[4696]: I0321 09:12:51.584353 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-526fx\" (UniqueName: \"kubernetes.io/projected/7d6364e4-44f2-42c2-b980-1ad5c1faba7b-kube-api-access-526fx\") pod \"redhat-operators-jbt2w\" (UID: \"7d6364e4-44f2-42c2-b980-1ad5c1faba7b\") " pod="openshift-marketplace/redhat-operators-jbt2w" Mar 21 09:12:51 crc kubenswrapper[4696]: I0321 09:12:51.584660 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d6364e4-44f2-42c2-b980-1ad5c1faba7b-utilities\") pod \"redhat-operators-jbt2w\" (UID: \"7d6364e4-44f2-42c2-b980-1ad5c1faba7b\") " pod="openshift-marketplace/redhat-operators-jbt2w" Mar 21 09:12:51 crc kubenswrapper[4696]: I0321 09:12:51.686327 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d6364e4-44f2-42c2-b980-1ad5c1faba7b-utilities\") pod \"redhat-operators-jbt2w\" (UID: \"7d6364e4-44f2-42c2-b980-1ad5c1faba7b\") " pod="openshift-marketplace/redhat-operators-jbt2w" Mar 21 09:12:51 crc kubenswrapper[4696]: I0321 09:12:51.686516 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d6364e4-44f2-42c2-b980-1ad5c1faba7b-catalog-content\") pod \"redhat-operators-jbt2w\" (UID: \"7d6364e4-44f2-42c2-b980-1ad5c1faba7b\") " pod="openshift-marketplace/redhat-operators-jbt2w" Mar 21 09:12:51 crc kubenswrapper[4696]: I0321 09:12:51.686549 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-526fx\" (UniqueName: \"kubernetes.io/projected/7d6364e4-44f2-42c2-b980-1ad5c1faba7b-kube-api-access-526fx\") pod \"redhat-operators-jbt2w\" (UID: \"7d6364e4-44f2-42c2-b980-1ad5c1faba7b\") " pod="openshift-marketplace/redhat-operators-jbt2w" Mar 21 09:12:51 crc kubenswrapper[4696]: I0321 09:12:51.686962 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d6364e4-44f2-42c2-b980-1ad5c1faba7b-utilities\") pod \"redhat-operators-jbt2w\" (UID: \"7d6364e4-44f2-42c2-b980-1ad5c1faba7b\") " pod="openshift-marketplace/redhat-operators-jbt2w" Mar 21 09:12:51 crc kubenswrapper[4696]: I0321 09:12:51.687188 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d6364e4-44f2-42c2-b980-1ad5c1faba7b-catalog-content\") pod \"redhat-operators-jbt2w\" (UID: \"7d6364e4-44f2-42c2-b980-1ad5c1faba7b\") " pod="openshift-marketplace/redhat-operators-jbt2w" Mar 21 09:12:51 crc kubenswrapper[4696]: I0321 09:12:51.706303 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-526fx\" (UniqueName: \"kubernetes.io/projected/7d6364e4-44f2-42c2-b980-1ad5c1faba7b-kube-api-access-526fx\") pod \"redhat-operators-jbt2w\" (UID: \"7d6364e4-44f2-42c2-b980-1ad5c1faba7b\") " pod="openshift-marketplace/redhat-operators-jbt2w" Mar 21 09:12:51 crc kubenswrapper[4696]: I0321 09:12:51.888058 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbt2w" Mar 21 09:12:52 crc kubenswrapper[4696]: I0321 09:12:52.473138 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jbt2w"] Mar 21 09:12:52 crc kubenswrapper[4696]: I0321 09:12:52.645988 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbt2w" event={"ID":"7d6364e4-44f2-42c2-b980-1ad5c1faba7b","Type":"ContainerStarted","Data":"490e98293d5e3072f0ab4d1a15c878f88ca955b339c55b5f5272741b9e5aa1a4"} Mar 21 09:12:53 crc kubenswrapper[4696]: I0321 09:12:53.657493 4696 generic.go:334] "Generic (PLEG): container finished" podID="7d6364e4-44f2-42c2-b980-1ad5c1faba7b" containerID="6b017ece245adcd868d368efb55c0640bdd4272e5ac40455e5c0e3a4f1776c6d" exitCode=0 Mar 21 09:12:53 crc kubenswrapper[4696]: I0321 09:12:53.657739 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbt2w" event={"ID":"7d6364e4-44f2-42c2-b980-1ad5c1faba7b","Type":"ContainerDied","Data":"6b017ece245adcd868d368efb55c0640bdd4272e5ac40455e5c0e3a4f1776c6d"} Mar 21 09:12:55 crc kubenswrapper[4696]: I0321 09:12:55.681361 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbt2w" event={"ID":"7d6364e4-44f2-42c2-b980-1ad5c1faba7b","Type":"ContainerStarted","Data":"a7d4a50685e6c5223f22ac1d8531d16d92ba83d4eec0781977ac7755fb7f622f"} Mar 21 09:12:59 crc kubenswrapper[4696]: I0321 09:12:59.730128 4696 generic.go:334] "Generic (PLEG): container finished" podID="7d6364e4-44f2-42c2-b980-1ad5c1faba7b" containerID="a7d4a50685e6c5223f22ac1d8531d16d92ba83d4eec0781977ac7755fb7f622f" exitCode=0 Mar 21 09:12:59 crc kubenswrapper[4696]: I0321 09:12:59.730256 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbt2w" event={"ID":"7d6364e4-44f2-42c2-b980-1ad5c1faba7b","Type":"ContainerDied","Data":"a7d4a50685e6c5223f22ac1d8531d16d92ba83d4eec0781977ac7755fb7f622f"} Mar 21 09:13:00 crc kubenswrapper[4696]: I0321 09:13:00.741313 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbt2w" event={"ID":"7d6364e4-44f2-42c2-b980-1ad5c1faba7b","Type":"ContainerStarted","Data":"12e915a75c56bd9192de2ed3e9b8002debcbe86f91fe23e49994de51f901cbcc"} Mar 21 09:13:00 crc kubenswrapper[4696]: I0321 09:13:00.763249 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jbt2w" podStartSLOduration=3.246741581 podStartE2EDuration="9.763231262s" podCreationTimestamp="2026-03-21 09:12:51 +0000 UTC" firstStartedPulling="2026-03-21 09:12:53.660400525 +0000 UTC m=+2707.781281238" lastFinishedPulling="2026-03-21 09:13:00.176890206 +0000 UTC m=+2714.297770919" observedRunningTime="2026-03-21 09:13:00.756184898 +0000 UTC m=+2714.877065611" watchObservedRunningTime="2026-03-21 09:13:00.763231262 +0000 UTC m=+2714.884111975" Mar 21 09:13:01 crc kubenswrapper[4696]: I0321 09:13:01.888864 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jbt2w" Mar 21 09:13:01 crc kubenswrapper[4696]: I0321 09:13:01.889158 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jbt2w" Mar 21 09:13:02 crc kubenswrapper[4696]: I0321 09:13:02.035760 4696 scope.go:117] "RemoveContainer" containerID="71c0fab430ec4d9914d39524313c6dbe3ae1f7aa18064eb07d884b0d68b6f3e7" Mar 21 09:13:02 crc kubenswrapper[4696]: I0321 09:13:02.933544 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jbt2w" podUID="7d6364e4-44f2-42c2-b980-1ad5c1faba7b" containerName="registry-server" probeResult="failure" output=< Mar 21 09:13:02 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Mar 21 09:13:02 crc kubenswrapper[4696]: > Mar 21 09:13:12 crc kubenswrapper[4696]: I0321 09:13:12.940154 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jbt2w" podUID="7d6364e4-44f2-42c2-b980-1ad5c1faba7b" containerName="registry-server" probeResult="failure" output=< Mar 21 09:13:12 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Mar 21 09:13:12 crc kubenswrapper[4696]: > Mar 21 09:13:22 crc kubenswrapper[4696]: I0321 09:13:22.945413 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jbt2w" podUID="7d6364e4-44f2-42c2-b980-1ad5c1faba7b" containerName="registry-server" probeResult="failure" output=< Mar 21 09:13:22 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Mar 21 09:13:22 crc kubenswrapper[4696]: > Mar 21 09:13:32 crc kubenswrapper[4696]: I0321 09:13:32.937052 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jbt2w" podUID="7d6364e4-44f2-42c2-b980-1ad5c1faba7b" containerName="registry-server" probeResult="failure" output=< Mar 21 09:13:32 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Mar 21 09:13:32 crc kubenswrapper[4696]: > Mar 21 09:13:34 crc kubenswrapper[4696]: E0321 09:13:34.491024 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Mar 21 09:13:34 crc kubenswrapper[4696]: E0321 09:13:34.491214 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kx624,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 09:13:34 crc kubenswrapper[4696]: E0321 09:13:34.492477 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e" Mar 21 09:13:35 crc kubenswrapper[4696]: E0321 09:13:35.078810 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e" Mar 21 09:13:42 crc kubenswrapper[4696]: I0321 09:13:42.953416 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jbt2w" podUID="7d6364e4-44f2-42c2-b980-1ad5c1faba7b" containerName="registry-server" probeResult="failure" output=< Mar 21 09:13:42 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Mar 21 09:13:42 crc kubenswrapper[4696]: > Mar 21 09:13:48 crc kubenswrapper[4696]: I0321 09:13:48.545449 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 09:13:49 crc kubenswrapper[4696]: I0321 09:13:49.112345 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Mar 21 09:13:50 crc kubenswrapper[4696]: I0321 09:13:50.218653 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e","Type":"ContainerStarted","Data":"21a940c2024b2ae674ffeeff8f9dbf4b76570261d812e607953ccb17a9d91226"} Mar 21 09:13:50 crc kubenswrapper[4696]: I0321 09:13:50.242208 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.668436235 podStartE2EDuration="1m12.242185868s" podCreationTimestamp="2026-03-21 09:12:38 +0000 UTC" firstStartedPulling="2026-03-21 09:12:40.531682904 +0000 UTC m=+2694.652563617" lastFinishedPulling="2026-03-21 09:13:49.105432537 +0000 UTC m=+2763.226313250" observedRunningTime="2026-03-21 09:13:50.233260342 +0000 UTC m=+2764.354141055" watchObservedRunningTime="2026-03-21 09:13:50.242185868 +0000 UTC m=+2764.363066581" Mar 21 09:13:51 crc kubenswrapper[4696]: I0321 09:13:51.938845 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jbt2w" Mar 21 09:13:51 crc kubenswrapper[4696]: I0321 09:13:51.991367 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jbt2w" Mar 21 09:13:52 crc kubenswrapper[4696]: I0321 09:13:52.735088 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jbt2w"] Mar 21 09:13:53 crc kubenswrapper[4696]: I0321 09:13:53.242650 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jbt2w" podUID="7d6364e4-44f2-42c2-b980-1ad5c1faba7b" containerName="registry-server" containerID="cri-o://12e915a75c56bd9192de2ed3e9b8002debcbe86f91fe23e49994de51f901cbcc" gracePeriod=2 Mar 21 09:13:53 crc kubenswrapper[4696]: I0321 09:13:53.865402 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbt2w" Mar 21 09:13:53 crc kubenswrapper[4696]: I0321 09:13:53.993649 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d6364e4-44f2-42c2-b980-1ad5c1faba7b-catalog-content\") pod \"7d6364e4-44f2-42c2-b980-1ad5c1faba7b\" (UID: \"7d6364e4-44f2-42c2-b980-1ad5c1faba7b\") " Mar 21 09:13:53 crc kubenswrapper[4696]: I0321 09:13:53.994141 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d6364e4-44f2-42c2-b980-1ad5c1faba7b-utilities\") pod \"7d6364e4-44f2-42c2-b980-1ad5c1faba7b\" (UID: \"7d6364e4-44f2-42c2-b980-1ad5c1faba7b\") " Mar 21 09:13:53 crc kubenswrapper[4696]: I0321 09:13:53.994264 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-526fx\" (UniqueName: \"kubernetes.io/projected/7d6364e4-44f2-42c2-b980-1ad5c1faba7b-kube-api-access-526fx\") pod \"7d6364e4-44f2-42c2-b980-1ad5c1faba7b\" (UID: \"7d6364e4-44f2-42c2-b980-1ad5c1faba7b\") " Mar 21 09:13:53 crc kubenswrapper[4696]: I0321 09:13:53.994522 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d6364e4-44f2-42c2-b980-1ad5c1faba7b-utilities" (OuterVolumeSpecName: "utilities") pod "7d6364e4-44f2-42c2-b980-1ad5c1faba7b" (UID: "7d6364e4-44f2-42c2-b980-1ad5c1faba7b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:13:53 crc kubenswrapper[4696]: I0321 09:13:53.995013 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d6364e4-44f2-42c2-b980-1ad5c1faba7b-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 09:13:54 crc kubenswrapper[4696]: I0321 09:13:54.004036 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d6364e4-44f2-42c2-b980-1ad5c1faba7b-kube-api-access-526fx" (OuterVolumeSpecName: "kube-api-access-526fx") pod "7d6364e4-44f2-42c2-b980-1ad5c1faba7b" (UID: "7d6364e4-44f2-42c2-b980-1ad5c1faba7b"). InnerVolumeSpecName "kube-api-access-526fx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:13:54 crc kubenswrapper[4696]: I0321 09:13:54.096953 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-526fx\" (UniqueName: \"kubernetes.io/projected/7d6364e4-44f2-42c2-b980-1ad5c1faba7b-kube-api-access-526fx\") on node \"crc\" DevicePath \"\"" Mar 21 09:13:54 crc kubenswrapper[4696]: I0321 09:13:54.113806 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d6364e4-44f2-42c2-b980-1ad5c1faba7b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d6364e4-44f2-42c2-b980-1ad5c1faba7b" (UID: "7d6364e4-44f2-42c2-b980-1ad5c1faba7b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:13:54 crc kubenswrapper[4696]: I0321 09:13:54.198724 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d6364e4-44f2-42c2-b980-1ad5c1faba7b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 09:13:54 crc kubenswrapper[4696]: I0321 09:13:54.257871 4696 generic.go:334] "Generic (PLEG): container finished" podID="7d6364e4-44f2-42c2-b980-1ad5c1faba7b" containerID="12e915a75c56bd9192de2ed3e9b8002debcbe86f91fe23e49994de51f901cbcc" exitCode=0 Mar 21 09:13:54 crc kubenswrapper[4696]: I0321 09:13:54.257913 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbt2w" event={"ID":"7d6364e4-44f2-42c2-b980-1ad5c1faba7b","Type":"ContainerDied","Data":"12e915a75c56bd9192de2ed3e9b8002debcbe86f91fe23e49994de51f901cbcc"} Mar 21 09:13:54 crc kubenswrapper[4696]: I0321 09:13:54.257942 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbt2w" event={"ID":"7d6364e4-44f2-42c2-b980-1ad5c1faba7b","Type":"ContainerDied","Data":"490e98293d5e3072f0ab4d1a15c878f88ca955b339c55b5f5272741b9e5aa1a4"} Mar 21 09:13:54 crc kubenswrapper[4696]: I0321 09:13:54.257960 4696 scope.go:117] "RemoveContainer" containerID="12e915a75c56bd9192de2ed3e9b8002debcbe86f91fe23e49994de51f901cbcc" Mar 21 09:13:54 crc kubenswrapper[4696]: I0321 09:13:54.257963 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbt2w" Mar 21 09:13:54 crc kubenswrapper[4696]: I0321 09:13:54.285501 4696 scope.go:117] "RemoveContainer" containerID="a7d4a50685e6c5223f22ac1d8531d16d92ba83d4eec0781977ac7755fb7f622f" Mar 21 09:13:54 crc kubenswrapper[4696]: I0321 09:13:54.293728 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jbt2w"] Mar 21 09:13:54 crc kubenswrapper[4696]: I0321 09:13:54.302930 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jbt2w"] Mar 21 09:13:54 crc kubenswrapper[4696]: I0321 09:13:54.316000 4696 scope.go:117] "RemoveContainer" containerID="6b017ece245adcd868d368efb55c0640bdd4272e5ac40455e5c0e3a4f1776c6d" Mar 21 09:13:54 crc kubenswrapper[4696]: I0321 09:13:54.360929 4696 scope.go:117] "RemoveContainer" containerID="12e915a75c56bd9192de2ed3e9b8002debcbe86f91fe23e49994de51f901cbcc" Mar 21 09:13:54 crc kubenswrapper[4696]: E0321 09:13:54.361398 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12e915a75c56bd9192de2ed3e9b8002debcbe86f91fe23e49994de51f901cbcc\": container with ID starting with 12e915a75c56bd9192de2ed3e9b8002debcbe86f91fe23e49994de51f901cbcc not found: ID does not exist" containerID="12e915a75c56bd9192de2ed3e9b8002debcbe86f91fe23e49994de51f901cbcc" Mar 21 09:13:54 crc kubenswrapper[4696]: I0321 09:13:54.361529 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12e915a75c56bd9192de2ed3e9b8002debcbe86f91fe23e49994de51f901cbcc"} err="failed to get container status \"12e915a75c56bd9192de2ed3e9b8002debcbe86f91fe23e49994de51f901cbcc\": rpc error: code = NotFound desc = could not find container \"12e915a75c56bd9192de2ed3e9b8002debcbe86f91fe23e49994de51f901cbcc\": container with ID starting with 12e915a75c56bd9192de2ed3e9b8002debcbe86f91fe23e49994de51f901cbcc not found: ID does not exist" Mar 21 09:13:54 crc kubenswrapper[4696]: I0321 09:13:54.361626 4696 scope.go:117] "RemoveContainer" containerID="a7d4a50685e6c5223f22ac1d8531d16d92ba83d4eec0781977ac7755fb7f622f" Mar 21 09:13:54 crc kubenswrapper[4696]: E0321 09:13:54.361940 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7d4a50685e6c5223f22ac1d8531d16d92ba83d4eec0781977ac7755fb7f622f\": container with ID starting with a7d4a50685e6c5223f22ac1d8531d16d92ba83d4eec0781977ac7755fb7f622f not found: ID does not exist" containerID="a7d4a50685e6c5223f22ac1d8531d16d92ba83d4eec0781977ac7755fb7f622f" Mar 21 09:13:54 crc kubenswrapper[4696]: I0321 09:13:54.361969 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7d4a50685e6c5223f22ac1d8531d16d92ba83d4eec0781977ac7755fb7f622f"} err="failed to get container status \"a7d4a50685e6c5223f22ac1d8531d16d92ba83d4eec0781977ac7755fb7f622f\": rpc error: code = NotFound desc = could not find container \"a7d4a50685e6c5223f22ac1d8531d16d92ba83d4eec0781977ac7755fb7f622f\": container with ID starting with a7d4a50685e6c5223f22ac1d8531d16d92ba83d4eec0781977ac7755fb7f622f not found: ID does not exist" Mar 21 09:13:54 crc kubenswrapper[4696]: I0321 09:13:54.361986 4696 scope.go:117] "RemoveContainer" containerID="6b017ece245adcd868d368efb55c0640bdd4272e5ac40455e5c0e3a4f1776c6d" Mar 21 09:13:54 crc kubenswrapper[4696]: E0321 09:13:54.362257 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b017ece245adcd868d368efb55c0640bdd4272e5ac40455e5c0e3a4f1776c6d\": container with ID starting with 6b017ece245adcd868d368efb55c0640bdd4272e5ac40455e5c0e3a4f1776c6d not found: ID does not exist" containerID="6b017ece245adcd868d368efb55c0640bdd4272e5ac40455e5c0e3a4f1776c6d" Mar 21 09:13:54 crc kubenswrapper[4696]: I0321 09:13:54.362349 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b017ece245adcd868d368efb55c0640bdd4272e5ac40455e5c0e3a4f1776c6d"} err="failed to get container status \"6b017ece245adcd868d368efb55c0640bdd4272e5ac40455e5c0e3a4f1776c6d\": rpc error: code = NotFound desc = could not find container \"6b017ece245adcd868d368efb55c0640bdd4272e5ac40455e5c0e3a4f1776c6d\": container with ID starting with 6b017ece245adcd868d368efb55c0640bdd4272e5ac40455e5c0e3a4f1776c6d not found: ID does not exist" Mar 21 09:13:54 crc kubenswrapper[4696]: I0321 09:13:54.546214 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d6364e4-44f2-42c2-b980-1ad5c1faba7b" path="/var/lib/kubelet/pods/7d6364e4-44f2-42c2-b980-1ad5c1faba7b/volumes" Mar 21 09:14:00 crc kubenswrapper[4696]: I0321 09:14:00.155747 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568074-rj4vk"] Mar 21 09:14:00 crc kubenswrapper[4696]: E0321 09:14:00.156676 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d6364e4-44f2-42c2-b980-1ad5c1faba7b" containerName="registry-server" Mar 21 09:14:00 crc kubenswrapper[4696]: I0321 09:14:00.156690 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d6364e4-44f2-42c2-b980-1ad5c1faba7b" containerName="registry-server" Mar 21 09:14:00 crc kubenswrapper[4696]: E0321 09:14:00.156708 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d6364e4-44f2-42c2-b980-1ad5c1faba7b" containerName="extract-utilities" Mar 21 09:14:00 crc kubenswrapper[4696]: I0321 09:14:00.156715 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d6364e4-44f2-42c2-b980-1ad5c1faba7b" containerName="extract-utilities" Mar 21 09:14:00 crc kubenswrapper[4696]: E0321 09:14:00.156743 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d6364e4-44f2-42c2-b980-1ad5c1faba7b" containerName="extract-content" Mar 21 09:14:00 crc kubenswrapper[4696]: I0321 09:14:00.156750 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d6364e4-44f2-42c2-b980-1ad5c1faba7b" containerName="extract-content" Mar 21 09:14:00 crc kubenswrapper[4696]: I0321 09:14:00.157002 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d6364e4-44f2-42c2-b980-1ad5c1faba7b" containerName="registry-server" Mar 21 09:14:00 crc kubenswrapper[4696]: I0321 09:14:00.157769 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568074-rj4vk" Mar 21 09:14:00 crc kubenswrapper[4696]: I0321 09:14:00.162037 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:14:00 crc kubenswrapper[4696]: I0321 09:14:00.162047 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:14:00 crc kubenswrapper[4696]: I0321 09:14:00.165752 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:14:00 crc kubenswrapper[4696]: I0321 09:14:00.170360 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568074-rj4vk"] Mar 21 09:14:00 crc kubenswrapper[4696]: I0321 09:14:00.232069 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r29gr\" (UniqueName: \"kubernetes.io/projected/d7a9ee89-ebf7-402e-bb18-a3037a3ffe19-kube-api-access-r29gr\") pod \"auto-csr-approver-29568074-rj4vk\" (UID: \"d7a9ee89-ebf7-402e-bb18-a3037a3ffe19\") " pod="openshift-infra/auto-csr-approver-29568074-rj4vk" Mar 21 09:14:00 crc kubenswrapper[4696]: I0321 09:14:00.334860 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r29gr\" (UniqueName: \"kubernetes.io/projected/d7a9ee89-ebf7-402e-bb18-a3037a3ffe19-kube-api-access-r29gr\") pod \"auto-csr-approver-29568074-rj4vk\" (UID: \"d7a9ee89-ebf7-402e-bb18-a3037a3ffe19\") " pod="openshift-infra/auto-csr-approver-29568074-rj4vk" Mar 21 09:14:00 crc kubenswrapper[4696]: I0321 09:14:00.341524 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:14:00 crc kubenswrapper[4696]: I0321 09:14:00.341580 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:14:00 crc kubenswrapper[4696]: I0321 09:14:00.352793 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r29gr\" (UniqueName: \"kubernetes.io/projected/d7a9ee89-ebf7-402e-bb18-a3037a3ffe19-kube-api-access-r29gr\") pod \"auto-csr-approver-29568074-rj4vk\" (UID: \"d7a9ee89-ebf7-402e-bb18-a3037a3ffe19\") " pod="openshift-infra/auto-csr-approver-29568074-rj4vk" Mar 21 09:14:00 crc kubenswrapper[4696]: I0321 09:14:00.478464 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568074-rj4vk" Mar 21 09:14:00 crc kubenswrapper[4696]: I0321 09:14:00.953172 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568074-rj4vk"] Mar 21 09:14:01 crc kubenswrapper[4696]: I0321 09:14:01.343492 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568074-rj4vk" event={"ID":"d7a9ee89-ebf7-402e-bb18-a3037a3ffe19","Type":"ContainerStarted","Data":"683f96c69c65dbf3562ec08d56afb098ccdb401fe84f1158854467e28ae3a450"} Mar 21 09:14:02 crc kubenswrapper[4696]: I0321 09:14:02.352938 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568074-rj4vk" event={"ID":"d7a9ee89-ebf7-402e-bb18-a3037a3ffe19","Type":"ContainerStarted","Data":"d00acf6f06b21d903f6d452196d15d49fee7e5c55bb07214c1af3274e676d356"} Mar 21 09:14:02 crc kubenswrapper[4696]: I0321 09:14:02.371285 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29568074-rj4vk" podStartSLOduration=1.544163459 podStartE2EDuration="2.371264923s" podCreationTimestamp="2026-03-21 09:14:00 +0000 UTC" firstStartedPulling="2026-03-21 09:14:00.961518206 +0000 UTC m=+2775.082398919" lastFinishedPulling="2026-03-21 09:14:01.78861967 +0000 UTC m=+2775.909500383" observedRunningTime="2026-03-21 09:14:02.367133999 +0000 UTC m=+2776.488014732" watchObservedRunningTime="2026-03-21 09:14:02.371264923 +0000 UTC m=+2776.492145636" Mar 21 09:14:03 crc kubenswrapper[4696]: I0321 09:14:03.376488 4696 generic.go:334] "Generic (PLEG): container finished" podID="d7a9ee89-ebf7-402e-bb18-a3037a3ffe19" containerID="d00acf6f06b21d903f6d452196d15d49fee7e5c55bb07214c1af3274e676d356" exitCode=0 Mar 21 09:14:03 crc kubenswrapper[4696]: I0321 09:14:03.376537 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568074-rj4vk" event={"ID":"d7a9ee89-ebf7-402e-bb18-a3037a3ffe19","Type":"ContainerDied","Data":"d00acf6f06b21d903f6d452196d15d49fee7e5c55bb07214c1af3274e676d356"} Mar 21 09:14:04 crc kubenswrapper[4696]: I0321 09:14:04.838715 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568074-rj4vk" Mar 21 09:14:04 crc kubenswrapper[4696]: I0321 09:14:04.928420 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r29gr\" (UniqueName: \"kubernetes.io/projected/d7a9ee89-ebf7-402e-bb18-a3037a3ffe19-kube-api-access-r29gr\") pod \"d7a9ee89-ebf7-402e-bb18-a3037a3ffe19\" (UID: \"d7a9ee89-ebf7-402e-bb18-a3037a3ffe19\") " Mar 21 09:14:04 crc kubenswrapper[4696]: I0321 09:14:04.933617 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7a9ee89-ebf7-402e-bb18-a3037a3ffe19-kube-api-access-r29gr" (OuterVolumeSpecName: "kube-api-access-r29gr") pod "d7a9ee89-ebf7-402e-bb18-a3037a3ffe19" (UID: "d7a9ee89-ebf7-402e-bb18-a3037a3ffe19"). InnerVolumeSpecName "kube-api-access-r29gr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:14:05 crc kubenswrapper[4696]: I0321 09:14:05.032035 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r29gr\" (UniqueName: \"kubernetes.io/projected/d7a9ee89-ebf7-402e-bb18-a3037a3ffe19-kube-api-access-r29gr\") on node \"crc\" DevicePath \"\"" Mar 21 09:14:05 crc kubenswrapper[4696]: I0321 09:14:05.398471 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568074-rj4vk" event={"ID":"d7a9ee89-ebf7-402e-bb18-a3037a3ffe19","Type":"ContainerDied","Data":"683f96c69c65dbf3562ec08d56afb098ccdb401fe84f1158854467e28ae3a450"} Mar 21 09:14:05 crc kubenswrapper[4696]: I0321 09:14:05.398776 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="683f96c69c65dbf3562ec08d56afb098ccdb401fe84f1158854467e28ae3a450" Mar 21 09:14:05 crc kubenswrapper[4696]: I0321 09:14:05.398513 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568074-rj4vk" Mar 21 09:14:05 crc kubenswrapper[4696]: I0321 09:14:05.453424 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568068-6f9cd"] Mar 21 09:14:05 crc kubenswrapper[4696]: I0321 09:14:05.480160 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568068-6f9cd"] Mar 21 09:14:06 crc kubenswrapper[4696]: I0321 09:14:06.549289 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bae18e6-4af5-467d-a736-78b4d04a0fa3" path="/var/lib/kubelet/pods/1bae18e6-4af5-467d-a736-78b4d04a0fa3/volumes" Mar 21 09:14:30 crc kubenswrapper[4696]: I0321 09:14:30.340967 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:14:30 crc kubenswrapper[4696]: I0321 09:14:30.341508 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.149218 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568075-rcmlh"] Mar 21 09:15:00 crc kubenswrapper[4696]: E0321 09:15:00.150964 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7a9ee89-ebf7-402e-bb18-a3037a3ffe19" containerName="oc" Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.151042 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7a9ee89-ebf7-402e-bb18-a3037a3ffe19" containerName="oc" Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.151326 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7a9ee89-ebf7-402e-bb18-a3037a3ffe19" containerName="oc" Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.152226 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568075-rcmlh" Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.154944 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.154996 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.157753 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568075-rcmlh"] Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.203794 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87sf4\" (UniqueName: \"kubernetes.io/projected/a203a166-f341-4d88-9ed0-3d540d770dec-kube-api-access-87sf4\") pod \"collect-profiles-29568075-rcmlh\" (UID: \"a203a166-f341-4d88-9ed0-3d540d770dec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568075-rcmlh" Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.204176 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a203a166-f341-4d88-9ed0-3d540d770dec-secret-volume\") pod \"collect-profiles-29568075-rcmlh\" (UID: \"a203a166-f341-4d88-9ed0-3d540d770dec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568075-rcmlh" Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.204257 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a203a166-f341-4d88-9ed0-3d540d770dec-config-volume\") pod \"collect-profiles-29568075-rcmlh\" (UID: \"a203a166-f341-4d88-9ed0-3d540d770dec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568075-rcmlh" Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.306223 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a203a166-f341-4d88-9ed0-3d540d770dec-config-volume\") pod \"collect-profiles-29568075-rcmlh\" (UID: \"a203a166-f341-4d88-9ed0-3d540d770dec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568075-rcmlh" Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.306561 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87sf4\" (UniqueName: \"kubernetes.io/projected/a203a166-f341-4d88-9ed0-3d540d770dec-kube-api-access-87sf4\") pod \"collect-profiles-29568075-rcmlh\" (UID: \"a203a166-f341-4d88-9ed0-3d540d770dec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568075-rcmlh" Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.306698 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a203a166-f341-4d88-9ed0-3d540d770dec-secret-volume\") pod \"collect-profiles-29568075-rcmlh\" (UID: \"a203a166-f341-4d88-9ed0-3d540d770dec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568075-rcmlh" Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.307319 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a203a166-f341-4d88-9ed0-3d540d770dec-config-volume\") pod \"collect-profiles-29568075-rcmlh\" (UID: \"a203a166-f341-4d88-9ed0-3d540d770dec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568075-rcmlh" Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.321480 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a203a166-f341-4d88-9ed0-3d540d770dec-secret-volume\") pod \"collect-profiles-29568075-rcmlh\" (UID: \"a203a166-f341-4d88-9ed0-3d540d770dec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568075-rcmlh" Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.328697 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87sf4\" (UniqueName: \"kubernetes.io/projected/a203a166-f341-4d88-9ed0-3d540d770dec-kube-api-access-87sf4\") pod \"collect-profiles-29568075-rcmlh\" (UID: \"a203a166-f341-4d88-9ed0-3d540d770dec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568075-rcmlh" Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.341392 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.341459 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.341512 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.342458 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a2f3653fbac2953f5711ffc6be0faa87b5c64ec53ea926ab1544f651d7adfcdd"} pod="openshift-machine-config-operator/machine-config-daemon-z7srw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.342526 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" containerID="cri-o://a2f3653fbac2953f5711ffc6be0faa87b5c64ec53ea926ab1544f651d7adfcdd" gracePeriod=600 Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.502398 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568075-rcmlh" Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.956255 4696 generic.go:334] "Generic (PLEG): container finished" podID="daaf227a-2305-495e-8495-a280abcd8e10" containerID="a2f3653fbac2953f5711ffc6be0faa87b5c64ec53ea926ab1544f651d7adfcdd" exitCode=0 Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.956344 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerDied","Data":"a2f3653fbac2953f5711ffc6be0faa87b5c64ec53ea926ab1544f651d7adfcdd"} Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.956883 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerStarted","Data":"5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c"} Mar 21 09:15:00 crc kubenswrapper[4696]: I0321 09:15:00.956914 4696 scope.go:117] "RemoveContainer" containerID="d3b5021774f92e92e24dfc72aea4cc0547d26a51f63f6954cb8360e1201f5263" Mar 21 09:15:01 crc kubenswrapper[4696]: I0321 09:15:01.147741 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568075-rcmlh"] Mar 21 09:15:01 crc kubenswrapper[4696]: W0321 09:15:01.149871 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda203a166_f341_4d88_9ed0_3d540d770dec.slice/crio-72cbd6b2056f52253fe50ce4489381d178134d32368ad18282f3d9cdb0a3e0f7 WatchSource:0}: Error finding container 72cbd6b2056f52253fe50ce4489381d178134d32368ad18282f3d9cdb0a3e0f7: Status 404 returned error can't find the container with id 72cbd6b2056f52253fe50ce4489381d178134d32368ad18282f3d9cdb0a3e0f7 Mar 21 09:15:02 crc kubenswrapper[4696]: I0321 09:15:02.000291 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568075-rcmlh" event={"ID":"a203a166-f341-4d88-9ed0-3d540d770dec","Type":"ContainerStarted","Data":"9c39d1544088817dbeb381c9737aad991f551cbd8ff6ac38949bf3b7a1de8869"} Mar 21 09:15:02 crc kubenswrapper[4696]: I0321 09:15:02.001071 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568075-rcmlh" event={"ID":"a203a166-f341-4d88-9ed0-3d540d770dec","Type":"ContainerStarted","Data":"72cbd6b2056f52253fe50ce4489381d178134d32368ad18282f3d9cdb0a3e0f7"} Mar 21 09:15:02 crc kubenswrapper[4696]: I0321 09:15:02.020030 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29568075-rcmlh" podStartSLOduration=2.020014556 podStartE2EDuration="2.020014556s" podCreationTimestamp="2026-03-21 09:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 09:15:02.015449419 +0000 UTC m=+2836.136330132" watchObservedRunningTime="2026-03-21 09:15:02.020014556 +0000 UTC m=+2836.140895269" Mar 21 09:15:02 crc kubenswrapper[4696]: I0321 09:15:02.160410 4696 scope.go:117] "RemoveContainer" containerID="baef8d13fe1ecc93a2f2c42df2804e723f27215d40de1d80fd02f6c1da4adc42" Mar 21 09:15:03 crc kubenswrapper[4696]: I0321 09:15:03.011060 4696 generic.go:334] "Generic (PLEG): container finished" podID="a203a166-f341-4d88-9ed0-3d540d770dec" containerID="9c39d1544088817dbeb381c9737aad991f551cbd8ff6ac38949bf3b7a1de8869" exitCode=0 Mar 21 09:15:03 crc kubenswrapper[4696]: I0321 09:15:03.011299 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568075-rcmlh" event={"ID":"a203a166-f341-4d88-9ed0-3d540d770dec","Type":"ContainerDied","Data":"9c39d1544088817dbeb381c9737aad991f551cbd8ff6ac38949bf3b7a1de8869"} Mar 21 09:15:04 crc kubenswrapper[4696]: I0321 09:15:04.548673 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568075-rcmlh" Mar 21 09:15:04 crc kubenswrapper[4696]: I0321 09:15:04.690225 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87sf4\" (UniqueName: \"kubernetes.io/projected/a203a166-f341-4d88-9ed0-3d540d770dec-kube-api-access-87sf4\") pod \"a203a166-f341-4d88-9ed0-3d540d770dec\" (UID: \"a203a166-f341-4d88-9ed0-3d540d770dec\") " Mar 21 09:15:04 crc kubenswrapper[4696]: I0321 09:15:04.690524 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a203a166-f341-4d88-9ed0-3d540d770dec-config-volume\") pod \"a203a166-f341-4d88-9ed0-3d540d770dec\" (UID: \"a203a166-f341-4d88-9ed0-3d540d770dec\") " Mar 21 09:15:04 crc kubenswrapper[4696]: I0321 09:15:04.691066 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a203a166-f341-4d88-9ed0-3d540d770dec-secret-volume\") pod \"a203a166-f341-4d88-9ed0-3d540d770dec\" (UID: \"a203a166-f341-4d88-9ed0-3d540d770dec\") " Mar 21 09:15:04 crc kubenswrapper[4696]: I0321 09:15:04.691132 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a203a166-f341-4d88-9ed0-3d540d770dec-config-volume" (OuterVolumeSpecName: "config-volume") pod "a203a166-f341-4d88-9ed0-3d540d770dec" (UID: "a203a166-f341-4d88-9ed0-3d540d770dec"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 09:15:04 crc kubenswrapper[4696]: I0321 09:15:04.693086 4696 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a203a166-f341-4d88-9ed0-3d540d770dec-config-volume\") on node \"crc\" DevicePath \"\"" Mar 21 09:15:04 crc kubenswrapper[4696]: I0321 09:15:04.697401 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a203a166-f341-4d88-9ed0-3d540d770dec-kube-api-access-87sf4" (OuterVolumeSpecName: "kube-api-access-87sf4") pod "a203a166-f341-4d88-9ed0-3d540d770dec" (UID: "a203a166-f341-4d88-9ed0-3d540d770dec"). InnerVolumeSpecName "kube-api-access-87sf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:15:04 crc kubenswrapper[4696]: I0321 09:15:04.706521 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a203a166-f341-4d88-9ed0-3d540d770dec-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a203a166-f341-4d88-9ed0-3d540d770dec" (UID: "a203a166-f341-4d88-9ed0-3d540d770dec"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:15:04 crc kubenswrapper[4696]: I0321 09:15:04.795191 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87sf4\" (UniqueName: \"kubernetes.io/projected/a203a166-f341-4d88-9ed0-3d540d770dec-kube-api-access-87sf4\") on node \"crc\" DevicePath \"\"" Mar 21 09:15:04 crc kubenswrapper[4696]: I0321 09:15:04.795238 4696 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a203a166-f341-4d88-9ed0-3d540d770dec-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 21 09:15:05 crc kubenswrapper[4696]: I0321 09:15:05.034852 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568075-rcmlh" event={"ID":"a203a166-f341-4d88-9ed0-3d540d770dec","Type":"ContainerDied","Data":"72cbd6b2056f52253fe50ce4489381d178134d32368ad18282f3d9cdb0a3e0f7"} Mar 21 09:15:05 crc kubenswrapper[4696]: I0321 09:15:05.034895 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72cbd6b2056f52253fe50ce4489381d178134d32368ad18282f3d9cdb0a3e0f7" Mar 21 09:15:05 crc kubenswrapper[4696]: I0321 09:15:05.034891 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568075-rcmlh" Mar 21 09:15:05 crc kubenswrapper[4696]: I0321 09:15:05.625911 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568030-n2gf7"] Mar 21 09:15:05 crc kubenswrapper[4696]: I0321 09:15:05.634562 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568030-n2gf7"] Mar 21 09:15:06 crc kubenswrapper[4696]: I0321 09:15:06.554543 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df64449e-f8aa-4e41-9b54-1dcc659d995c" path="/var/lib/kubelet/pods/df64449e-f8aa-4e41-9b54-1dcc659d995c/volumes" Mar 21 09:16:00 crc kubenswrapper[4696]: I0321 09:16:00.152937 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568076-pchmg"] Mar 21 09:16:00 crc kubenswrapper[4696]: E0321 09:16:00.154274 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a203a166-f341-4d88-9ed0-3d540d770dec" containerName="collect-profiles" Mar 21 09:16:00 crc kubenswrapper[4696]: I0321 09:16:00.154296 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="a203a166-f341-4d88-9ed0-3d540d770dec" containerName="collect-profiles" Mar 21 09:16:00 crc kubenswrapper[4696]: I0321 09:16:00.154688 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="a203a166-f341-4d88-9ed0-3d540d770dec" containerName="collect-profiles" Mar 21 09:16:00 crc kubenswrapper[4696]: I0321 09:16:00.156032 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568076-pchmg" Mar 21 09:16:00 crc kubenswrapper[4696]: I0321 09:16:00.160124 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:16:00 crc kubenswrapper[4696]: I0321 09:16:00.160246 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:16:00 crc kubenswrapper[4696]: I0321 09:16:00.163648 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568076-pchmg"] Mar 21 09:16:00 crc kubenswrapper[4696]: I0321 09:16:00.169930 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:16:00 crc kubenswrapper[4696]: I0321 09:16:00.289474 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmgff\" (UniqueName: \"kubernetes.io/projected/20f13cbd-092d-4a84-97b5-3bb0b48b8e4d-kube-api-access-qmgff\") pod \"auto-csr-approver-29568076-pchmg\" (UID: \"20f13cbd-092d-4a84-97b5-3bb0b48b8e4d\") " pod="openshift-infra/auto-csr-approver-29568076-pchmg" Mar 21 09:16:00 crc kubenswrapper[4696]: I0321 09:16:00.392454 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmgff\" (UniqueName: \"kubernetes.io/projected/20f13cbd-092d-4a84-97b5-3bb0b48b8e4d-kube-api-access-qmgff\") pod \"auto-csr-approver-29568076-pchmg\" (UID: \"20f13cbd-092d-4a84-97b5-3bb0b48b8e4d\") " pod="openshift-infra/auto-csr-approver-29568076-pchmg" Mar 21 09:16:00 crc kubenswrapper[4696]: I0321 09:16:00.419554 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmgff\" (UniqueName: \"kubernetes.io/projected/20f13cbd-092d-4a84-97b5-3bb0b48b8e4d-kube-api-access-qmgff\") pod \"auto-csr-approver-29568076-pchmg\" (UID: \"20f13cbd-092d-4a84-97b5-3bb0b48b8e4d\") " pod="openshift-infra/auto-csr-approver-29568076-pchmg" Mar 21 09:16:00 crc kubenswrapper[4696]: I0321 09:16:00.474006 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568076-pchmg" Mar 21 09:16:00 crc kubenswrapper[4696]: I0321 09:16:00.968449 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568076-pchmg"] Mar 21 09:16:01 crc kubenswrapper[4696]: I0321 09:16:01.532915 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568076-pchmg" event={"ID":"20f13cbd-092d-4a84-97b5-3bb0b48b8e4d","Type":"ContainerStarted","Data":"d2c019c3ba8dcce533edbb94e5c145054bd01b4f04e5f258ed2d170ff9664bde"} Mar 21 09:16:02 crc kubenswrapper[4696]: I0321 09:16:02.231619 4696 scope.go:117] "RemoveContainer" containerID="85895d626856496826a9a8a44f692f0479e3f5cb33877cf1fccfb8cdc3892201" Mar 21 09:16:02 crc kubenswrapper[4696]: I0321 09:16:02.548391 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568076-pchmg" event={"ID":"20f13cbd-092d-4a84-97b5-3bb0b48b8e4d","Type":"ContainerStarted","Data":"bbd36ab2f3383ee30dc688a02c06ad325d4689455afd4fc7483648a434349e41"} Mar 21 09:16:02 crc kubenswrapper[4696]: I0321 09:16:02.567584 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29568076-pchmg" podStartSLOduration=1.6692982029999999 podStartE2EDuration="2.567565136s" podCreationTimestamp="2026-03-21 09:16:00 +0000 UTC" firstStartedPulling="2026-03-21 09:16:00.972128776 +0000 UTC m=+2895.093009489" lastFinishedPulling="2026-03-21 09:16:01.870395709 +0000 UTC m=+2895.991276422" observedRunningTime="2026-03-21 09:16:02.560571404 +0000 UTC m=+2896.681452117" watchObservedRunningTime="2026-03-21 09:16:02.567565136 +0000 UTC m=+2896.688445849" Mar 21 09:16:03 crc kubenswrapper[4696]: I0321 09:16:03.558509 4696 generic.go:334] "Generic (PLEG): container finished" podID="20f13cbd-092d-4a84-97b5-3bb0b48b8e4d" containerID="bbd36ab2f3383ee30dc688a02c06ad325d4689455afd4fc7483648a434349e41" exitCode=0 Mar 21 09:16:03 crc kubenswrapper[4696]: I0321 09:16:03.558573 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568076-pchmg" event={"ID":"20f13cbd-092d-4a84-97b5-3bb0b48b8e4d","Type":"ContainerDied","Data":"bbd36ab2f3383ee30dc688a02c06ad325d4689455afd4fc7483648a434349e41"} Mar 21 09:16:05 crc kubenswrapper[4696]: I0321 09:16:05.141169 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568076-pchmg" Mar 21 09:16:05 crc kubenswrapper[4696]: I0321 09:16:05.296600 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmgff\" (UniqueName: \"kubernetes.io/projected/20f13cbd-092d-4a84-97b5-3bb0b48b8e4d-kube-api-access-qmgff\") pod \"20f13cbd-092d-4a84-97b5-3bb0b48b8e4d\" (UID: \"20f13cbd-092d-4a84-97b5-3bb0b48b8e4d\") " Mar 21 09:16:05 crc kubenswrapper[4696]: I0321 09:16:05.313069 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20f13cbd-092d-4a84-97b5-3bb0b48b8e4d-kube-api-access-qmgff" (OuterVolumeSpecName: "kube-api-access-qmgff") pod "20f13cbd-092d-4a84-97b5-3bb0b48b8e4d" (UID: "20f13cbd-092d-4a84-97b5-3bb0b48b8e4d"). InnerVolumeSpecName "kube-api-access-qmgff". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:16:05 crc kubenswrapper[4696]: I0321 09:16:05.399140 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmgff\" (UniqueName: \"kubernetes.io/projected/20f13cbd-092d-4a84-97b5-3bb0b48b8e4d-kube-api-access-qmgff\") on node \"crc\" DevicePath \"\"" Mar 21 09:16:05 crc kubenswrapper[4696]: I0321 09:16:05.610154 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568076-pchmg" event={"ID":"20f13cbd-092d-4a84-97b5-3bb0b48b8e4d","Type":"ContainerDied","Data":"d2c019c3ba8dcce533edbb94e5c145054bd01b4f04e5f258ed2d170ff9664bde"} Mar 21 09:16:05 crc kubenswrapper[4696]: I0321 09:16:05.610213 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2c019c3ba8dcce533edbb94e5c145054bd01b4f04e5f258ed2d170ff9664bde" Mar 21 09:16:05 crc kubenswrapper[4696]: I0321 09:16:05.610316 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568076-pchmg" Mar 21 09:16:05 crc kubenswrapper[4696]: I0321 09:16:05.655891 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568070-sp48z"] Mar 21 09:16:05 crc kubenswrapper[4696]: I0321 09:16:05.665870 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568070-sp48z"] Mar 21 09:16:06 crc kubenswrapper[4696]: I0321 09:16:06.548778 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="303575eb-eb75-4bf4-8501-5e67a2c72470" path="/var/lib/kubelet/pods/303575eb-eb75-4bf4-8501-5e67a2c72470/volumes" Mar 21 09:17:00 crc kubenswrapper[4696]: I0321 09:17:00.341007 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:17:00 crc kubenswrapper[4696]: I0321 09:17:00.341569 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:17:02 crc kubenswrapper[4696]: I0321 09:17:02.326501 4696 scope.go:117] "RemoveContainer" containerID="1dd24cefb0132751f500df4ae16247679726ae6ff291d82371f5ed45d48b3d7f" Mar 21 09:17:30 crc kubenswrapper[4696]: I0321 09:17:30.341582 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:17:30 crc kubenswrapper[4696]: I0321 09:17:30.342025 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:18:00 crc kubenswrapper[4696]: I0321 09:18:00.150301 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568078-gprd7"] Mar 21 09:18:00 crc kubenswrapper[4696]: E0321 09:18:00.151245 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20f13cbd-092d-4a84-97b5-3bb0b48b8e4d" containerName="oc" Mar 21 09:18:00 crc kubenswrapper[4696]: I0321 09:18:00.151260 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="20f13cbd-092d-4a84-97b5-3bb0b48b8e4d" containerName="oc" Mar 21 09:18:00 crc kubenswrapper[4696]: I0321 09:18:00.151492 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="20f13cbd-092d-4a84-97b5-3bb0b48b8e4d" containerName="oc" Mar 21 09:18:00 crc kubenswrapper[4696]: I0321 09:18:00.152318 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568078-gprd7" Mar 21 09:18:00 crc kubenswrapper[4696]: I0321 09:18:00.154196 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:18:00 crc kubenswrapper[4696]: I0321 09:18:00.154317 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:18:00 crc kubenswrapper[4696]: I0321 09:18:00.154389 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:18:00 crc kubenswrapper[4696]: I0321 09:18:00.168697 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568078-gprd7"] Mar 21 09:18:00 crc kubenswrapper[4696]: I0321 09:18:00.262716 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxsxd\" (UniqueName: \"kubernetes.io/projected/07bdef47-c372-4b8f-a093-3a31c5231fb9-kube-api-access-vxsxd\") pod \"auto-csr-approver-29568078-gprd7\" (UID: \"07bdef47-c372-4b8f-a093-3a31c5231fb9\") " pod="openshift-infra/auto-csr-approver-29568078-gprd7" Mar 21 09:18:00 crc kubenswrapper[4696]: I0321 09:18:00.341667 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:18:00 crc kubenswrapper[4696]: I0321 09:18:00.341714 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:18:00 crc kubenswrapper[4696]: I0321 09:18:00.341767 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 09:18:00 crc kubenswrapper[4696]: I0321 09:18:00.342479 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c"} pod="openshift-machine-config-operator/machine-config-daemon-z7srw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 09:18:00 crc kubenswrapper[4696]: I0321 09:18:00.342533 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" containerID="cri-o://5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" gracePeriod=600 Mar 21 09:18:00 crc kubenswrapper[4696]: I0321 09:18:00.365137 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxsxd\" (UniqueName: \"kubernetes.io/projected/07bdef47-c372-4b8f-a093-3a31c5231fb9-kube-api-access-vxsxd\") pod \"auto-csr-approver-29568078-gprd7\" (UID: \"07bdef47-c372-4b8f-a093-3a31c5231fb9\") " pod="openshift-infra/auto-csr-approver-29568078-gprd7" Mar 21 09:18:00 crc kubenswrapper[4696]: I0321 09:18:00.389842 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxsxd\" (UniqueName: \"kubernetes.io/projected/07bdef47-c372-4b8f-a093-3a31c5231fb9-kube-api-access-vxsxd\") pod \"auto-csr-approver-29568078-gprd7\" (UID: \"07bdef47-c372-4b8f-a093-3a31c5231fb9\") " pod="openshift-infra/auto-csr-approver-29568078-gprd7" Mar 21 09:18:00 crc kubenswrapper[4696]: I0321 09:18:00.472631 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568078-gprd7" Mar 21 09:18:00 crc kubenswrapper[4696]: E0321 09:18:00.503412 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:18:00 crc kubenswrapper[4696]: I0321 09:18:00.700733 4696 generic.go:334] "Generic (PLEG): container finished" podID="daaf227a-2305-495e-8495-a280abcd8e10" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" exitCode=0 Mar 21 09:18:00 crc kubenswrapper[4696]: I0321 09:18:00.700776 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerDied","Data":"5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c"} Mar 21 09:18:00 crc kubenswrapper[4696]: I0321 09:18:00.700809 4696 scope.go:117] "RemoveContainer" containerID="a2f3653fbac2953f5711ffc6be0faa87b5c64ec53ea926ab1544f651d7adfcdd" Mar 21 09:18:00 crc kubenswrapper[4696]: I0321 09:18:00.701608 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:18:00 crc kubenswrapper[4696]: E0321 09:18:00.701869 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:18:01 crc kubenswrapper[4696]: I0321 09:18:01.075527 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568078-gprd7"] Mar 21 09:18:01 crc kubenswrapper[4696]: I0321 09:18:01.714602 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568078-gprd7" event={"ID":"07bdef47-c372-4b8f-a093-3a31c5231fb9","Type":"ContainerStarted","Data":"e174f0131277211d96c65058701143170e99f5c99fb24f85d9a9cd556c1f6ee7"} Mar 21 09:18:02 crc kubenswrapper[4696]: I0321 09:18:02.725672 4696 generic.go:334] "Generic (PLEG): container finished" podID="07bdef47-c372-4b8f-a093-3a31c5231fb9" containerID="11da25fec5cad12f50d5cd8a74c193bf96237eb669b12b759bc6a941e076c8f5" exitCode=0 Mar 21 09:18:02 crc kubenswrapper[4696]: I0321 09:18:02.725776 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568078-gprd7" event={"ID":"07bdef47-c372-4b8f-a093-3a31c5231fb9","Type":"ContainerDied","Data":"11da25fec5cad12f50d5cd8a74c193bf96237eb669b12b759bc6a941e076c8f5"} Mar 21 09:18:04 crc kubenswrapper[4696]: I0321 09:18:04.546740 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568078-gprd7" Mar 21 09:18:04 crc kubenswrapper[4696]: I0321 09:18:04.562081 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxsxd\" (UniqueName: \"kubernetes.io/projected/07bdef47-c372-4b8f-a093-3a31c5231fb9-kube-api-access-vxsxd\") pod \"07bdef47-c372-4b8f-a093-3a31c5231fb9\" (UID: \"07bdef47-c372-4b8f-a093-3a31c5231fb9\") " Mar 21 09:18:04 crc kubenswrapper[4696]: I0321 09:18:04.568000 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07bdef47-c372-4b8f-a093-3a31c5231fb9-kube-api-access-vxsxd" (OuterVolumeSpecName: "kube-api-access-vxsxd") pod "07bdef47-c372-4b8f-a093-3a31c5231fb9" (UID: "07bdef47-c372-4b8f-a093-3a31c5231fb9"). InnerVolumeSpecName "kube-api-access-vxsxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:18:04 crc kubenswrapper[4696]: I0321 09:18:04.664719 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxsxd\" (UniqueName: \"kubernetes.io/projected/07bdef47-c372-4b8f-a093-3a31c5231fb9-kube-api-access-vxsxd\") on node \"crc\" DevicePath \"\"" Mar 21 09:18:04 crc kubenswrapper[4696]: I0321 09:18:04.743541 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568078-gprd7" event={"ID":"07bdef47-c372-4b8f-a093-3a31c5231fb9","Type":"ContainerDied","Data":"e174f0131277211d96c65058701143170e99f5c99fb24f85d9a9cd556c1f6ee7"} Mar 21 09:18:04 crc kubenswrapper[4696]: I0321 09:18:04.743582 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e174f0131277211d96c65058701143170e99f5c99fb24f85d9a9cd556c1f6ee7" Mar 21 09:18:04 crc kubenswrapper[4696]: I0321 09:18:04.743605 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568078-gprd7" Mar 21 09:18:05 crc kubenswrapper[4696]: I0321 09:18:05.638647 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568072-nkw88"] Mar 21 09:18:05 crc kubenswrapper[4696]: I0321 09:18:05.647779 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568072-nkw88"] Mar 21 09:18:06 crc kubenswrapper[4696]: I0321 09:18:06.545234 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09438c8e-71b8-4bdb-a32e-e07d0d641b0a" path="/var/lib/kubelet/pods/09438c8e-71b8-4bdb-a32e-e07d0d641b0a/volumes" Mar 21 09:18:12 crc kubenswrapper[4696]: I0321 09:18:12.534334 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:18:12 crc kubenswrapper[4696]: E0321 09:18:12.535198 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:18:27 crc kubenswrapper[4696]: I0321 09:18:27.535915 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:18:27 crc kubenswrapper[4696]: E0321 09:18:27.536625 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:18:38 crc kubenswrapper[4696]: I0321 09:18:38.535905 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:18:38 crc kubenswrapper[4696]: E0321 09:18:38.536693 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:18:51 crc kubenswrapper[4696]: I0321 09:18:51.534341 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:18:51 crc kubenswrapper[4696]: E0321 09:18:51.535207 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:19:02 crc kubenswrapper[4696]: I0321 09:19:02.440067 4696 scope.go:117] "RemoveContainer" containerID="94f55f343f50a788ca0600e9a5903a4bdd94d7990d1db08b260ef92786b722ee" Mar 21 09:19:05 crc kubenswrapper[4696]: I0321 09:19:05.534782 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:19:05 crc kubenswrapper[4696]: E0321 09:19:05.535552 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:19:20 crc kubenswrapper[4696]: I0321 09:19:20.535776 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:19:20 crc kubenswrapper[4696]: E0321 09:19:20.536614 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:19:28 crc kubenswrapper[4696]: I0321 09:19:28.546895 4696 generic.go:334] "Generic (PLEG): container finished" podID="1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e" containerID="21a940c2024b2ae674ffeeff8f9dbf4b76570261d812e607953ccb17a9d91226" exitCode=0 Mar 21 09:19:28 crc kubenswrapper[4696]: I0321 09:19:28.555458 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e","Type":"ContainerDied","Data":"21a940c2024b2ae674ffeeff8f9dbf4b76570261d812e607953ccb17a9d91226"} Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.579618 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e","Type":"ContainerDied","Data":"4820203f47462f9c4ceaaa5022751be45b3923440b83ad12d76a1afcdf34a416"} Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.580167 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4820203f47462f9c4ceaaa5022751be45b3923440b83ad12d76a1afcdf34a416" Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.588694 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.629910 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kx624\" (UniqueName: \"kubernetes.io/projected/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-kube-api-access-kx624\") pod \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.630003 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-test-operator-ephemeral-workdir\") pod \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.630109 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.630155 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-config-data\") pod \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.630238 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-openstack-config\") pod \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.630312 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-ca-certs\") pod \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.630397 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-openstack-config-secret\") pod \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.630421 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-ssh-key\") pod \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.630521 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-test-operator-ephemeral-temporary\") pod \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\" (UID: \"1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e\") " Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.637598 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-kube-api-access-kx624" (OuterVolumeSpecName: "kube-api-access-kx624") pod "1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e" (UID: "1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e"). InnerVolumeSpecName "kube-api-access-kx624". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.638618 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e" (UID: "1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.646620 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-config-data" (OuterVolumeSpecName: "config-data") pod "1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e" (UID: "1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.660634 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e" (UID: "1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.674971 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e" (UID: "1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.680759 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e" (UID: "1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.696998 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e" (UID: "1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.717647 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e" (UID: "1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.734100 4696 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.734132 4696 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-ssh-key\") on node \"crc\" DevicePath \"\"" Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.734143 4696 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.734154 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kx624\" (UniqueName: \"kubernetes.io/projected/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-kube-api-access-kx624\") on node \"crc\" DevicePath \"\"" Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.734177 4696 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.734189 4696 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-config-data\") on node \"crc\" DevicePath \"\"" Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.734199 4696 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.734207 4696 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-ca-certs\") on node \"crc\" DevicePath \"\"" Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.775276 4696 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Mar 21 09:19:30 crc kubenswrapper[4696]: I0321 09:19:30.838650 4696 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Mar 21 09:19:31 crc kubenswrapper[4696]: I0321 09:19:31.131086 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e" (UID: "1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:19:31 crc kubenswrapper[4696]: I0321 09:19:31.144307 4696 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Mar 21 09:19:31 crc kubenswrapper[4696]: I0321 09:19:31.587506 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 21 09:19:34 crc kubenswrapper[4696]: I0321 09:19:34.535447 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:19:34 crc kubenswrapper[4696]: E0321 09:19:34.536163 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:19:38 crc kubenswrapper[4696]: I0321 09:19:38.433357 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 21 09:19:38 crc kubenswrapper[4696]: E0321 09:19:38.434431 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07bdef47-c372-4b8f-a093-3a31c5231fb9" containerName="oc" Mar 21 09:19:38 crc kubenswrapper[4696]: I0321 09:19:38.434448 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="07bdef47-c372-4b8f-a093-3a31c5231fb9" containerName="oc" Mar 21 09:19:38 crc kubenswrapper[4696]: E0321 09:19:38.434490 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e" containerName="tempest-tests-tempest-tests-runner" Mar 21 09:19:38 crc kubenswrapper[4696]: I0321 09:19:38.434499 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e" containerName="tempest-tests-tempest-tests-runner" Mar 21 09:19:38 crc kubenswrapper[4696]: I0321 09:19:38.434897 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e" containerName="tempest-tests-tempest-tests-runner" Mar 21 09:19:38 crc kubenswrapper[4696]: I0321 09:19:38.434965 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="07bdef47-c372-4b8f-a093-3a31c5231fb9" containerName="oc" Mar 21 09:19:38 crc kubenswrapper[4696]: I0321 09:19:38.437190 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 21 09:19:38 crc kubenswrapper[4696]: I0321 09:19:38.441385 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-cdj78" Mar 21 09:19:38 crc kubenswrapper[4696]: I0321 09:19:38.467261 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 21 09:19:38 crc kubenswrapper[4696]: I0321 09:19:38.598201 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e6c3f461-61e5-443c-911d-0b8a1ba72a96\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 21 09:19:38 crc kubenswrapper[4696]: I0321 09:19:38.598654 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmxwv\" (UniqueName: \"kubernetes.io/projected/e6c3f461-61e5-443c-911d-0b8a1ba72a96-kube-api-access-fmxwv\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e6c3f461-61e5-443c-911d-0b8a1ba72a96\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 21 09:19:38 crc kubenswrapper[4696]: I0321 09:19:38.701647 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e6c3f461-61e5-443c-911d-0b8a1ba72a96\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 21 09:19:38 crc kubenswrapper[4696]: I0321 09:19:38.716403 4696 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e6c3f461-61e5-443c-911d-0b8a1ba72a96\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 21 09:19:38 crc kubenswrapper[4696]: I0321 09:19:38.742808 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmxwv\" (UniqueName: \"kubernetes.io/projected/e6c3f461-61e5-443c-911d-0b8a1ba72a96-kube-api-access-fmxwv\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e6c3f461-61e5-443c-911d-0b8a1ba72a96\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 21 09:19:38 crc kubenswrapper[4696]: I0321 09:19:38.782409 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmxwv\" (UniqueName: \"kubernetes.io/projected/e6c3f461-61e5-443c-911d-0b8a1ba72a96-kube-api-access-fmxwv\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e6c3f461-61e5-443c-911d-0b8a1ba72a96\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 21 09:19:38 crc kubenswrapper[4696]: I0321 09:19:38.787622 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e6c3f461-61e5-443c-911d-0b8a1ba72a96\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 21 09:19:39 crc kubenswrapper[4696]: I0321 09:19:39.088239 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 21 09:19:39 crc kubenswrapper[4696]: I0321 09:19:39.827083 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 21 09:19:39 crc kubenswrapper[4696]: I0321 09:19:39.834151 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 09:19:40 crc kubenswrapper[4696]: I0321 09:19:40.711018 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"e6c3f461-61e5-443c-911d-0b8a1ba72a96","Type":"ContainerStarted","Data":"db718905424147922afea5a17d6829cbb0d13dd3f856170c28629e92a6da6222"} Mar 21 09:19:41 crc kubenswrapper[4696]: I0321 09:19:41.721923 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"e6c3f461-61e5-443c-911d-0b8a1ba72a96","Type":"ContainerStarted","Data":"0fbb7545dc601c5202f36d0d96688a9bc622f00acfc6bfe0a46b60ac24519e7d"} Mar 21 09:19:41 crc kubenswrapper[4696]: I0321 09:19:41.765454 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.701321237 podStartE2EDuration="3.765430558s" podCreationTimestamp="2026-03-21 09:19:38 +0000 UTC" firstStartedPulling="2026-03-21 09:19:39.833939585 +0000 UTC m=+3113.954820298" lastFinishedPulling="2026-03-21 09:19:40.898048906 +0000 UTC m=+3115.018929619" observedRunningTime="2026-03-21 09:19:41.7575466 +0000 UTC m=+3115.878427313" watchObservedRunningTime="2026-03-21 09:19:41.765430558 +0000 UTC m=+3115.886311271" Mar 21 09:19:49 crc kubenswrapper[4696]: I0321 09:19:49.534835 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:19:49 crc kubenswrapper[4696]: E0321 09:19:49.535589 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:20:00 crc kubenswrapper[4696]: I0321 09:20:00.178713 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568080-82ltv"] Mar 21 09:20:00 crc kubenswrapper[4696]: I0321 09:20:00.180675 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568080-82ltv" Mar 21 09:20:00 crc kubenswrapper[4696]: I0321 09:20:00.184541 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:20:00 crc kubenswrapper[4696]: I0321 09:20:00.185117 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:20:00 crc kubenswrapper[4696]: I0321 09:20:00.186736 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:20:00 crc kubenswrapper[4696]: I0321 09:20:00.191839 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568080-82ltv"] Mar 21 09:20:00 crc kubenswrapper[4696]: I0321 09:20:00.330348 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dlgt\" (UniqueName: \"kubernetes.io/projected/640729b7-7edd-4af6-b156-811c9aaf7574-kube-api-access-5dlgt\") pod \"auto-csr-approver-29568080-82ltv\" (UID: \"640729b7-7edd-4af6-b156-811c9aaf7574\") " pod="openshift-infra/auto-csr-approver-29568080-82ltv" Mar 21 09:20:00 crc kubenswrapper[4696]: I0321 09:20:00.432438 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dlgt\" (UniqueName: \"kubernetes.io/projected/640729b7-7edd-4af6-b156-811c9aaf7574-kube-api-access-5dlgt\") pod \"auto-csr-approver-29568080-82ltv\" (UID: \"640729b7-7edd-4af6-b156-811c9aaf7574\") " pod="openshift-infra/auto-csr-approver-29568080-82ltv" Mar 21 09:20:00 crc kubenswrapper[4696]: I0321 09:20:00.455971 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dlgt\" (UniqueName: \"kubernetes.io/projected/640729b7-7edd-4af6-b156-811c9aaf7574-kube-api-access-5dlgt\") pod \"auto-csr-approver-29568080-82ltv\" (UID: \"640729b7-7edd-4af6-b156-811c9aaf7574\") " pod="openshift-infra/auto-csr-approver-29568080-82ltv" Mar 21 09:20:00 crc kubenswrapper[4696]: I0321 09:20:00.524727 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568080-82ltv" Mar 21 09:20:00 crc kubenswrapper[4696]: I0321 09:20:00.535632 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:20:00 crc kubenswrapper[4696]: E0321 09:20:00.535937 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:20:01 crc kubenswrapper[4696]: I0321 09:20:01.267534 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568080-82ltv"] Mar 21 09:20:01 crc kubenswrapper[4696]: I0321 09:20:01.903277 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568080-82ltv" event={"ID":"640729b7-7edd-4af6-b156-811c9aaf7574","Type":"ContainerStarted","Data":"9c6bc77df054e89388171569f4847d26e1f6fd98fd6f235ea4d7b621b85f1413"} Mar 21 09:20:02 crc kubenswrapper[4696]: I0321 09:20:02.907429 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mckgv/must-gather-kqwf9"] Mar 21 09:20:02 crc kubenswrapper[4696]: I0321 09:20:02.909462 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mckgv/must-gather-kqwf9" Mar 21 09:20:02 crc kubenswrapper[4696]: I0321 09:20:02.916649 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568080-82ltv" event={"ID":"640729b7-7edd-4af6-b156-811c9aaf7574","Type":"ContainerStarted","Data":"0835b4d56b56640ee07e7dda24d19c240cf0efb8d3a9ff0b21bc3ea9e413778e"} Mar 21 09:20:02 crc kubenswrapper[4696]: I0321 09:20:02.919925 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mckgv"/"openshift-service-ca.crt" Mar 21 09:20:02 crc kubenswrapper[4696]: I0321 09:20:02.920421 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mckgv"/"kube-root-ca.crt" Mar 21 09:20:02 crc kubenswrapper[4696]: I0321 09:20:02.932004 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mckgv/must-gather-kqwf9"] Mar 21 09:20:02 crc kubenswrapper[4696]: I0321 09:20:02.979428 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29568080-82ltv" podStartSLOduration=1.89234955 podStartE2EDuration="2.979411377s" podCreationTimestamp="2026-03-21 09:20:00 +0000 UTC" firstStartedPulling="2026-03-21 09:20:01.278259344 +0000 UTC m=+3135.399140057" lastFinishedPulling="2026-03-21 09:20:02.365321171 +0000 UTC m=+3136.486201884" observedRunningTime="2026-03-21 09:20:02.974083999 +0000 UTC m=+3137.094964712" watchObservedRunningTime="2026-03-21 09:20:02.979411377 +0000 UTC m=+3137.100292090" Mar 21 09:20:02 crc kubenswrapper[4696]: I0321 09:20:02.990978 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6d6e8c8f-6a59-443a-9874-848be975ebd1-must-gather-output\") pod \"must-gather-kqwf9\" (UID: \"6d6e8c8f-6a59-443a-9874-848be975ebd1\") " pod="openshift-must-gather-mckgv/must-gather-kqwf9" Mar 21 09:20:02 crc kubenswrapper[4696]: I0321 09:20:02.991144 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7wp6\" (UniqueName: \"kubernetes.io/projected/6d6e8c8f-6a59-443a-9874-848be975ebd1-kube-api-access-q7wp6\") pod \"must-gather-kqwf9\" (UID: \"6d6e8c8f-6a59-443a-9874-848be975ebd1\") " pod="openshift-must-gather-mckgv/must-gather-kqwf9" Mar 21 09:20:03 crc kubenswrapper[4696]: I0321 09:20:03.093385 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6d6e8c8f-6a59-443a-9874-848be975ebd1-must-gather-output\") pod \"must-gather-kqwf9\" (UID: \"6d6e8c8f-6a59-443a-9874-848be975ebd1\") " pod="openshift-must-gather-mckgv/must-gather-kqwf9" Mar 21 09:20:03 crc kubenswrapper[4696]: I0321 09:20:03.093554 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7wp6\" (UniqueName: \"kubernetes.io/projected/6d6e8c8f-6a59-443a-9874-848be975ebd1-kube-api-access-q7wp6\") pod \"must-gather-kqwf9\" (UID: \"6d6e8c8f-6a59-443a-9874-848be975ebd1\") " pod="openshift-must-gather-mckgv/must-gather-kqwf9" Mar 21 09:20:03 crc kubenswrapper[4696]: I0321 09:20:03.094146 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6d6e8c8f-6a59-443a-9874-848be975ebd1-must-gather-output\") pod \"must-gather-kqwf9\" (UID: \"6d6e8c8f-6a59-443a-9874-848be975ebd1\") " pod="openshift-must-gather-mckgv/must-gather-kqwf9" Mar 21 09:20:03 crc kubenswrapper[4696]: I0321 09:20:03.121317 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7wp6\" (UniqueName: \"kubernetes.io/projected/6d6e8c8f-6a59-443a-9874-848be975ebd1-kube-api-access-q7wp6\") pod \"must-gather-kqwf9\" (UID: \"6d6e8c8f-6a59-443a-9874-848be975ebd1\") " pod="openshift-must-gather-mckgv/must-gather-kqwf9" Mar 21 09:20:03 crc kubenswrapper[4696]: I0321 09:20:03.225679 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mckgv/must-gather-kqwf9" Mar 21 09:20:03 crc kubenswrapper[4696]: I0321 09:20:03.928218 4696 generic.go:334] "Generic (PLEG): container finished" podID="640729b7-7edd-4af6-b156-811c9aaf7574" containerID="0835b4d56b56640ee07e7dda24d19c240cf0efb8d3a9ff0b21bc3ea9e413778e" exitCode=0 Mar 21 09:20:03 crc kubenswrapper[4696]: I0321 09:20:03.928287 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568080-82ltv" event={"ID":"640729b7-7edd-4af6-b156-811c9aaf7574","Type":"ContainerDied","Data":"0835b4d56b56640ee07e7dda24d19c240cf0efb8d3a9ff0b21bc3ea9e413778e"} Mar 21 09:20:04 crc kubenswrapper[4696]: W0321 09:20:04.012808 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d6e8c8f_6a59_443a_9874_848be975ebd1.slice/crio-852ac132d56eec7b20b47fb7a069d9bca49dadcde41d6e78dcf0c76662be5bb9 WatchSource:0}: Error finding container 852ac132d56eec7b20b47fb7a069d9bca49dadcde41d6e78dcf0c76662be5bb9: Status 404 returned error can't find the container with id 852ac132d56eec7b20b47fb7a069d9bca49dadcde41d6e78dcf0c76662be5bb9 Mar 21 09:20:04 crc kubenswrapper[4696]: I0321 09:20:04.013099 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mckgv/must-gather-kqwf9"] Mar 21 09:20:04 crc kubenswrapper[4696]: I0321 09:20:04.940746 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mckgv/must-gather-kqwf9" event={"ID":"6d6e8c8f-6a59-443a-9874-848be975ebd1","Type":"ContainerStarted","Data":"852ac132d56eec7b20b47fb7a069d9bca49dadcde41d6e78dcf0c76662be5bb9"} Mar 21 09:20:06 crc kubenswrapper[4696]: I0321 09:20:06.097925 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568080-82ltv" Mar 21 09:20:06 crc kubenswrapper[4696]: I0321 09:20:06.161771 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dlgt\" (UniqueName: \"kubernetes.io/projected/640729b7-7edd-4af6-b156-811c9aaf7574-kube-api-access-5dlgt\") pod \"640729b7-7edd-4af6-b156-811c9aaf7574\" (UID: \"640729b7-7edd-4af6-b156-811c9aaf7574\") " Mar 21 09:20:06 crc kubenswrapper[4696]: I0321 09:20:06.169056 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/640729b7-7edd-4af6-b156-811c9aaf7574-kube-api-access-5dlgt" (OuterVolumeSpecName: "kube-api-access-5dlgt") pod "640729b7-7edd-4af6-b156-811c9aaf7574" (UID: "640729b7-7edd-4af6-b156-811c9aaf7574"). InnerVolumeSpecName "kube-api-access-5dlgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:20:06 crc kubenswrapper[4696]: I0321 09:20:06.264056 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dlgt\" (UniqueName: \"kubernetes.io/projected/640729b7-7edd-4af6-b156-811c9aaf7574-kube-api-access-5dlgt\") on node \"crc\" DevicePath \"\"" Mar 21 09:20:06 crc kubenswrapper[4696]: I0321 09:20:06.959377 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568080-82ltv" event={"ID":"640729b7-7edd-4af6-b156-811c9aaf7574","Type":"ContainerDied","Data":"9c6bc77df054e89388171569f4847d26e1f6fd98fd6f235ea4d7b621b85f1413"} Mar 21 09:20:06 crc kubenswrapper[4696]: I0321 09:20:06.959413 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c6bc77df054e89388171569f4847d26e1f6fd98fd6f235ea4d7b621b85f1413" Mar 21 09:20:06 crc kubenswrapper[4696]: I0321 09:20:06.959461 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568080-82ltv" Mar 21 09:20:07 crc kubenswrapper[4696]: I0321 09:20:07.180017 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568074-rj4vk"] Mar 21 09:20:07 crc kubenswrapper[4696]: I0321 09:20:07.193402 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568074-rj4vk"] Mar 21 09:20:08 crc kubenswrapper[4696]: I0321 09:20:08.550178 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7a9ee89-ebf7-402e-bb18-a3037a3ffe19" path="/var/lib/kubelet/pods/d7a9ee89-ebf7-402e-bb18-a3037a3ffe19/volumes" Mar 21 09:20:12 crc kubenswrapper[4696]: I0321 09:20:12.057532 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mckgv/must-gather-kqwf9" event={"ID":"6d6e8c8f-6a59-443a-9874-848be975ebd1","Type":"ContainerStarted","Data":"a1511802f4c23a501730dbcfa0f1b4f0babb12187a6dd728849dc817af142b12"} Mar 21 09:20:13 crc kubenswrapper[4696]: I0321 09:20:13.068037 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mckgv/must-gather-kqwf9" event={"ID":"6d6e8c8f-6a59-443a-9874-848be975ebd1","Type":"ContainerStarted","Data":"846c1d1531ac135a588967c945e79fda57f16e40397e1a85b1155ffd22a198e2"} Mar 21 09:20:13 crc kubenswrapper[4696]: I0321 09:20:13.088068 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mckgv/must-gather-kqwf9" podStartSLOduration=3.386921464 podStartE2EDuration="11.088046007s" podCreationTimestamp="2026-03-21 09:20:02 +0000 UTC" firstStartedPulling="2026-03-21 09:20:04.014835314 +0000 UTC m=+3138.135716027" lastFinishedPulling="2026-03-21 09:20:11.715959857 +0000 UTC m=+3145.836840570" observedRunningTime="2026-03-21 09:20:13.082190765 +0000 UTC m=+3147.203071478" watchObservedRunningTime="2026-03-21 09:20:13.088046007 +0000 UTC m=+3147.208926740" Mar 21 09:20:13 crc kubenswrapper[4696]: I0321 09:20:13.534969 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:20:13 crc kubenswrapper[4696]: E0321 09:20:13.535196 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:20:17 crc kubenswrapper[4696]: I0321 09:20:17.255230 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mckgv/crc-debug-98kw2"] Mar 21 09:20:17 crc kubenswrapper[4696]: E0321 09:20:17.256212 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="640729b7-7edd-4af6-b156-811c9aaf7574" containerName="oc" Mar 21 09:20:17 crc kubenswrapper[4696]: I0321 09:20:17.256229 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="640729b7-7edd-4af6-b156-811c9aaf7574" containerName="oc" Mar 21 09:20:17 crc kubenswrapper[4696]: I0321 09:20:17.256464 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="640729b7-7edd-4af6-b156-811c9aaf7574" containerName="oc" Mar 21 09:20:17 crc kubenswrapper[4696]: I0321 09:20:17.258077 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mckgv/crc-debug-98kw2" Mar 21 09:20:17 crc kubenswrapper[4696]: I0321 09:20:17.260001 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mckgv"/"default-dockercfg-fkd4t" Mar 21 09:20:17 crc kubenswrapper[4696]: I0321 09:20:17.310895 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7876d0a1-82db-4d12-9b2b-e85160810659-host\") pod \"crc-debug-98kw2\" (UID: \"7876d0a1-82db-4d12-9b2b-e85160810659\") " pod="openshift-must-gather-mckgv/crc-debug-98kw2" Mar 21 09:20:17 crc kubenswrapper[4696]: I0321 09:20:17.311160 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqqkb\" (UniqueName: \"kubernetes.io/projected/7876d0a1-82db-4d12-9b2b-e85160810659-kube-api-access-rqqkb\") pod \"crc-debug-98kw2\" (UID: \"7876d0a1-82db-4d12-9b2b-e85160810659\") " pod="openshift-must-gather-mckgv/crc-debug-98kw2" Mar 21 09:20:17 crc kubenswrapper[4696]: I0321 09:20:17.413420 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqqkb\" (UniqueName: \"kubernetes.io/projected/7876d0a1-82db-4d12-9b2b-e85160810659-kube-api-access-rqqkb\") pod \"crc-debug-98kw2\" (UID: \"7876d0a1-82db-4d12-9b2b-e85160810659\") " pod="openshift-must-gather-mckgv/crc-debug-98kw2" Mar 21 09:20:17 crc kubenswrapper[4696]: I0321 09:20:17.413613 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7876d0a1-82db-4d12-9b2b-e85160810659-host\") pod \"crc-debug-98kw2\" (UID: \"7876d0a1-82db-4d12-9b2b-e85160810659\") " pod="openshift-must-gather-mckgv/crc-debug-98kw2" Mar 21 09:20:17 crc kubenswrapper[4696]: I0321 09:20:17.413726 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7876d0a1-82db-4d12-9b2b-e85160810659-host\") pod \"crc-debug-98kw2\" (UID: \"7876d0a1-82db-4d12-9b2b-e85160810659\") " pod="openshift-must-gather-mckgv/crc-debug-98kw2" Mar 21 09:20:17 crc kubenswrapper[4696]: I0321 09:20:17.442471 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqqkb\" (UniqueName: \"kubernetes.io/projected/7876d0a1-82db-4d12-9b2b-e85160810659-kube-api-access-rqqkb\") pod \"crc-debug-98kw2\" (UID: \"7876d0a1-82db-4d12-9b2b-e85160810659\") " pod="openshift-must-gather-mckgv/crc-debug-98kw2" Mar 21 09:20:17 crc kubenswrapper[4696]: I0321 09:20:17.582087 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mckgv/crc-debug-98kw2" Mar 21 09:20:18 crc kubenswrapper[4696]: I0321 09:20:18.131432 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mckgv/crc-debug-98kw2" event={"ID":"7876d0a1-82db-4d12-9b2b-e85160810659","Type":"ContainerStarted","Data":"8f6a323f13c842e1a609442d3b35e1105c088ffb04031b67d356964bc142ac7d"} Mar 21 09:20:22 crc kubenswrapper[4696]: I0321 09:20:22.089856 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-llhxl"] Mar 21 09:20:22 crc kubenswrapper[4696]: I0321 09:20:22.092520 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-llhxl" Mar 21 09:20:22 crc kubenswrapper[4696]: I0321 09:20:22.107926 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-llhxl"] Mar 21 09:20:22 crc kubenswrapper[4696]: I0321 09:20:22.237729 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjfj4\" (UniqueName: \"kubernetes.io/projected/b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577-kube-api-access-wjfj4\") pod \"community-operators-llhxl\" (UID: \"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577\") " pod="openshift-marketplace/community-operators-llhxl" Mar 21 09:20:22 crc kubenswrapper[4696]: I0321 09:20:22.237919 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577-catalog-content\") pod \"community-operators-llhxl\" (UID: \"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577\") " pod="openshift-marketplace/community-operators-llhxl" Mar 21 09:20:22 crc kubenswrapper[4696]: I0321 09:20:22.237999 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577-utilities\") pod \"community-operators-llhxl\" (UID: \"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577\") " pod="openshift-marketplace/community-operators-llhxl" Mar 21 09:20:22 crc kubenswrapper[4696]: I0321 09:20:22.340157 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjfj4\" (UniqueName: \"kubernetes.io/projected/b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577-kube-api-access-wjfj4\") pod \"community-operators-llhxl\" (UID: \"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577\") " pod="openshift-marketplace/community-operators-llhxl" Mar 21 09:20:22 crc kubenswrapper[4696]: I0321 09:20:22.340306 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577-catalog-content\") pod \"community-operators-llhxl\" (UID: \"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577\") " pod="openshift-marketplace/community-operators-llhxl" Mar 21 09:20:22 crc kubenswrapper[4696]: I0321 09:20:22.340369 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577-utilities\") pod \"community-operators-llhxl\" (UID: \"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577\") " pod="openshift-marketplace/community-operators-llhxl" Mar 21 09:20:22 crc kubenswrapper[4696]: I0321 09:20:22.340952 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577-catalog-content\") pod \"community-operators-llhxl\" (UID: \"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577\") " pod="openshift-marketplace/community-operators-llhxl" Mar 21 09:20:22 crc kubenswrapper[4696]: I0321 09:20:22.340985 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577-utilities\") pod \"community-operators-llhxl\" (UID: \"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577\") " pod="openshift-marketplace/community-operators-llhxl" Mar 21 09:20:22 crc kubenswrapper[4696]: I0321 09:20:22.362878 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjfj4\" (UniqueName: \"kubernetes.io/projected/b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577-kube-api-access-wjfj4\") pod \"community-operators-llhxl\" (UID: \"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577\") " pod="openshift-marketplace/community-operators-llhxl" Mar 21 09:20:22 crc kubenswrapper[4696]: I0321 09:20:22.458211 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-llhxl" Mar 21 09:20:23 crc kubenswrapper[4696]: I0321 09:20:23.433068 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-llhxl"] Mar 21 09:20:24 crc kubenswrapper[4696]: I0321 09:20:24.208600 4696 generic.go:334] "Generic (PLEG): container finished" podID="b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577" containerID="dbd7bbcfe2966981d19b01763b96e71570a2a577f2c9f0251e1917a23e958d8d" exitCode=0 Mar 21 09:20:24 crc kubenswrapper[4696]: I0321 09:20:24.208682 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llhxl" event={"ID":"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577","Type":"ContainerDied","Data":"dbd7bbcfe2966981d19b01763b96e71570a2a577f2c9f0251e1917a23e958d8d"} Mar 21 09:20:24 crc kubenswrapper[4696]: I0321 09:20:24.208907 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llhxl" event={"ID":"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577","Type":"ContainerStarted","Data":"6830a2caaad9be7ad8dd7bb4edd602d5c27574f9e9a1176a0ec4649ffdb89227"} Mar 21 09:20:24 crc kubenswrapper[4696]: I0321 09:20:24.535083 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:20:24 crc kubenswrapper[4696]: E0321 09:20:24.535639 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:20:25 crc kubenswrapper[4696]: I0321 09:20:25.220046 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llhxl" event={"ID":"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577","Type":"ContainerStarted","Data":"5552fa86addffcf875b5faf94bc60294dafe7fd8051398cf77718d34b5abfa7c"} Mar 21 09:20:28 crc kubenswrapper[4696]: I0321 09:20:28.252543 4696 generic.go:334] "Generic (PLEG): container finished" podID="b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577" containerID="5552fa86addffcf875b5faf94bc60294dafe7fd8051398cf77718d34b5abfa7c" exitCode=0 Mar 21 09:20:28 crc kubenswrapper[4696]: I0321 09:20:28.252562 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llhxl" event={"ID":"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577","Type":"ContainerDied","Data":"5552fa86addffcf875b5faf94bc60294dafe7fd8051398cf77718d34b5abfa7c"} Mar 21 09:20:37 crc kubenswrapper[4696]: E0321 09:20:37.403963 4696 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296" Mar 21 09:20:37 crc kubenswrapper[4696]: E0321 09:20:37.405642 4696 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:container-00,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296,Command:[chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ; rm -rf \"/var/tmp/sos-osp\" && mkdir -p \"/var/tmp/sos-osp\" && sudo podman rm --force toolbox-osp; sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools && toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=\"/var/tmp/sos-osp\" && if [[ \"$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)\" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf \"/var/tmp/sos-osp/podlogs.tar.xz\" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TMOUT,Value:900,ValueFrom:nil,},EnvVar{Name:HOST,Value:/host,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host,ReadOnly:false,MountPath:/host,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rqqkb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod crc-debug-98kw2_openshift-must-gather-mckgv(7876d0a1-82db-4d12-9b2b-e85160810659): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 21 09:20:37 crc kubenswrapper[4696]: E0321 09:20:37.407061 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openshift-must-gather-mckgv/crc-debug-98kw2" podUID="7876d0a1-82db-4d12-9b2b-e85160810659" Mar 21 09:20:38 crc kubenswrapper[4696]: I0321 09:20:38.359216 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llhxl" event={"ID":"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577","Type":"ContainerStarted","Data":"001222a7defaec84e313a0bd3e488d4cdc0915c154a97ec30d18ca95fb5c7b11"} Mar 21 09:20:38 crc kubenswrapper[4696]: E0321 09:20:38.361117 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296\\\"\"" pod="openshift-must-gather-mckgv/crc-debug-98kw2" podUID="7876d0a1-82db-4d12-9b2b-e85160810659" Mar 21 09:20:39 crc kubenswrapper[4696]: I0321 09:20:39.535229 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:20:39 crc kubenswrapper[4696]: E0321 09:20:39.535852 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:20:42 crc kubenswrapper[4696]: I0321 09:20:42.458871 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-llhxl" Mar 21 09:20:42 crc kubenswrapper[4696]: I0321 09:20:42.462268 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-llhxl" Mar 21 09:20:42 crc kubenswrapper[4696]: I0321 09:20:42.764216 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-llhxl" Mar 21 09:20:42 crc kubenswrapper[4696]: I0321 09:20:42.799096 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-llhxl" podStartSLOduration=7.104987996 podStartE2EDuration="20.799076492s" podCreationTimestamp="2026-03-21 09:20:22 +0000 UTC" firstStartedPulling="2026-03-21 09:20:24.210604276 +0000 UTC m=+3158.331484989" lastFinishedPulling="2026-03-21 09:20:37.904692772 +0000 UTC m=+3172.025573485" observedRunningTime="2026-03-21 09:20:38.422551654 +0000 UTC m=+3172.543432377" watchObservedRunningTime="2026-03-21 09:20:42.799076492 +0000 UTC m=+3176.919957195" Mar 21 09:20:43 crc kubenswrapper[4696]: I0321 09:20:43.460434 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-llhxl" Mar 21 09:20:43 crc kubenswrapper[4696]: I0321 09:20:43.509648 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-llhxl"] Mar 21 09:20:45 crc kubenswrapper[4696]: I0321 09:20:45.422448 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-llhxl" podUID="b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577" containerName="registry-server" containerID="cri-o://001222a7defaec84e313a0bd3e488d4cdc0915c154a97ec30d18ca95fb5c7b11" gracePeriod=2 Mar 21 09:20:46 crc kubenswrapper[4696]: I0321 09:20:46.435629 4696 generic.go:334] "Generic (PLEG): container finished" podID="b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577" containerID="001222a7defaec84e313a0bd3e488d4cdc0915c154a97ec30d18ca95fb5c7b11" exitCode=0 Mar 21 09:20:46 crc kubenswrapper[4696]: I0321 09:20:46.435708 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llhxl" event={"ID":"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577","Type":"ContainerDied","Data":"001222a7defaec84e313a0bd3e488d4cdc0915c154a97ec30d18ca95fb5c7b11"} Mar 21 09:20:46 crc kubenswrapper[4696]: I0321 09:20:46.630207 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-llhxl" Mar 21 09:20:46 crc kubenswrapper[4696]: I0321 09:20:46.786904 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjfj4\" (UniqueName: \"kubernetes.io/projected/b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577-kube-api-access-wjfj4\") pod \"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577\" (UID: \"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577\") " Mar 21 09:20:46 crc kubenswrapper[4696]: I0321 09:20:46.786976 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577-catalog-content\") pod \"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577\" (UID: \"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577\") " Mar 21 09:20:46 crc kubenswrapper[4696]: I0321 09:20:46.787057 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577-utilities\") pod \"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577\" (UID: \"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577\") " Mar 21 09:20:46 crc kubenswrapper[4696]: I0321 09:20:46.788006 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577-utilities" (OuterVolumeSpecName: "utilities") pod "b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577" (UID: "b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:20:46 crc kubenswrapper[4696]: I0321 09:20:46.793221 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577-kube-api-access-wjfj4" (OuterVolumeSpecName: "kube-api-access-wjfj4") pod "b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577" (UID: "b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577"). InnerVolumeSpecName "kube-api-access-wjfj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:20:46 crc kubenswrapper[4696]: I0321 09:20:46.839428 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577" (UID: "b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:20:46 crc kubenswrapper[4696]: I0321 09:20:46.889277 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjfj4\" (UniqueName: \"kubernetes.io/projected/b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577-kube-api-access-wjfj4\") on node \"crc\" DevicePath \"\"" Mar 21 09:20:46 crc kubenswrapper[4696]: I0321 09:20:46.889601 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 09:20:46 crc kubenswrapper[4696]: I0321 09:20:46.889615 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 09:20:47 crc kubenswrapper[4696]: I0321 09:20:47.456136 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llhxl" event={"ID":"b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577","Type":"ContainerDied","Data":"6830a2caaad9be7ad8dd7bb4edd602d5c27574f9e9a1176a0ec4649ffdb89227"} Mar 21 09:20:47 crc kubenswrapper[4696]: I0321 09:20:47.456194 4696 scope.go:117] "RemoveContainer" containerID="001222a7defaec84e313a0bd3e488d4cdc0915c154a97ec30d18ca95fb5c7b11" Mar 21 09:20:47 crc kubenswrapper[4696]: I0321 09:20:47.456361 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-llhxl" Mar 21 09:20:47 crc kubenswrapper[4696]: I0321 09:20:47.477189 4696 scope.go:117] "RemoveContainer" containerID="5552fa86addffcf875b5faf94bc60294dafe7fd8051398cf77718d34b5abfa7c" Mar 21 09:20:47 crc kubenswrapper[4696]: I0321 09:20:47.494296 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-llhxl"] Mar 21 09:20:47 crc kubenswrapper[4696]: I0321 09:20:47.508399 4696 scope.go:117] "RemoveContainer" containerID="dbd7bbcfe2966981d19b01763b96e71570a2a577f2c9f0251e1917a23e958d8d" Mar 21 09:20:47 crc kubenswrapper[4696]: I0321 09:20:47.510566 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-llhxl"] Mar 21 09:20:48 crc kubenswrapper[4696]: I0321 09:20:48.546284 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577" path="/var/lib/kubelet/pods/b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577/volumes" Mar 21 09:20:51 crc kubenswrapper[4696]: I0321 09:20:51.497531 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mckgv/crc-debug-98kw2" event={"ID":"7876d0a1-82db-4d12-9b2b-e85160810659","Type":"ContainerStarted","Data":"45f96776833c901506595cd02bf3295549f989f0d7816b89ae9800023766da5a"} Mar 21 09:20:51 crc kubenswrapper[4696]: I0321 09:20:51.528457 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mckgv/crc-debug-98kw2" podStartSLOduration=1.169870417 podStartE2EDuration="34.528438011s" podCreationTimestamp="2026-03-21 09:20:17 +0000 UTC" firstStartedPulling="2026-03-21 09:20:17.628243008 +0000 UTC m=+3151.749123711" lastFinishedPulling="2026-03-21 09:20:50.986810592 +0000 UTC m=+3185.107691305" observedRunningTime="2026-03-21 09:20:51.519766401 +0000 UTC m=+3185.640647114" watchObservedRunningTime="2026-03-21 09:20:51.528438011 +0000 UTC m=+3185.649318724" Mar 21 09:20:53 crc kubenswrapper[4696]: I0321 09:20:53.535332 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:20:53 crc kubenswrapper[4696]: E0321 09:20:53.536162 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:21:02 crc kubenswrapper[4696]: I0321 09:21:02.644873 4696 scope.go:117] "RemoveContainer" containerID="d00acf6f06b21d903f6d452196d15d49fee7e5c55bb07214c1af3274e676d356" Mar 21 09:21:06 crc kubenswrapper[4696]: I0321 09:21:06.543719 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:21:06 crc kubenswrapper[4696]: E0321 09:21:06.544411 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:21:17 crc kubenswrapper[4696]: I0321 09:21:17.534716 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:21:17 crc kubenswrapper[4696]: E0321 09:21:17.535439 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:21:31 crc kubenswrapper[4696]: I0321 09:21:31.534892 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:21:31 crc kubenswrapper[4696]: E0321 09:21:31.535782 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:21:45 crc kubenswrapper[4696]: I0321 09:21:45.535473 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:21:45 crc kubenswrapper[4696]: E0321 09:21:45.536104 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:21:49 crc kubenswrapper[4696]: I0321 09:21:49.079453 4696 generic.go:334] "Generic (PLEG): container finished" podID="7876d0a1-82db-4d12-9b2b-e85160810659" containerID="45f96776833c901506595cd02bf3295549f989f0d7816b89ae9800023766da5a" exitCode=0 Mar 21 09:21:49 crc kubenswrapper[4696]: I0321 09:21:49.079514 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mckgv/crc-debug-98kw2" event={"ID":"7876d0a1-82db-4d12-9b2b-e85160810659","Type":"ContainerDied","Data":"45f96776833c901506595cd02bf3295549f989f0d7816b89ae9800023766da5a"} Mar 21 09:21:50 crc kubenswrapper[4696]: I0321 09:21:50.229523 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mckgv/crc-debug-98kw2" Mar 21 09:21:50 crc kubenswrapper[4696]: I0321 09:21:50.261857 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mckgv/crc-debug-98kw2"] Mar 21 09:21:50 crc kubenswrapper[4696]: I0321 09:21:50.271102 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mckgv/crc-debug-98kw2"] Mar 21 09:21:50 crc kubenswrapper[4696]: I0321 09:21:50.332765 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7876d0a1-82db-4d12-9b2b-e85160810659-host\") pod \"7876d0a1-82db-4d12-9b2b-e85160810659\" (UID: \"7876d0a1-82db-4d12-9b2b-e85160810659\") " Mar 21 09:21:50 crc kubenswrapper[4696]: I0321 09:21:50.333036 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqqkb\" (UniqueName: \"kubernetes.io/projected/7876d0a1-82db-4d12-9b2b-e85160810659-kube-api-access-rqqkb\") pod \"7876d0a1-82db-4d12-9b2b-e85160810659\" (UID: \"7876d0a1-82db-4d12-9b2b-e85160810659\") " Mar 21 09:21:50 crc kubenswrapper[4696]: I0321 09:21:50.333251 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7876d0a1-82db-4d12-9b2b-e85160810659-host" (OuterVolumeSpecName: "host") pod "7876d0a1-82db-4d12-9b2b-e85160810659" (UID: "7876d0a1-82db-4d12-9b2b-e85160810659"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 09:21:50 crc kubenswrapper[4696]: I0321 09:21:50.333887 4696 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7876d0a1-82db-4d12-9b2b-e85160810659-host\") on node \"crc\" DevicePath \"\"" Mar 21 09:21:50 crc kubenswrapper[4696]: I0321 09:21:50.337764 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7876d0a1-82db-4d12-9b2b-e85160810659-kube-api-access-rqqkb" (OuterVolumeSpecName: "kube-api-access-rqqkb") pod "7876d0a1-82db-4d12-9b2b-e85160810659" (UID: "7876d0a1-82db-4d12-9b2b-e85160810659"). InnerVolumeSpecName "kube-api-access-rqqkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:21:50 crc kubenswrapper[4696]: I0321 09:21:50.435780 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqqkb\" (UniqueName: \"kubernetes.io/projected/7876d0a1-82db-4d12-9b2b-e85160810659-kube-api-access-rqqkb\") on node \"crc\" DevicePath \"\"" Mar 21 09:21:50 crc kubenswrapper[4696]: I0321 09:21:50.549754 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7876d0a1-82db-4d12-9b2b-e85160810659" path="/var/lib/kubelet/pods/7876d0a1-82db-4d12-9b2b-e85160810659/volumes" Mar 21 09:21:51 crc kubenswrapper[4696]: I0321 09:21:51.109307 4696 scope.go:117] "RemoveContainer" containerID="45f96776833c901506595cd02bf3295549f989f0d7816b89ae9800023766da5a" Mar 21 09:21:51 crc kubenswrapper[4696]: I0321 09:21:51.109555 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mckgv/crc-debug-98kw2" Mar 21 09:21:51 crc kubenswrapper[4696]: I0321 09:21:51.601034 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mckgv/crc-debug-rq9hw"] Mar 21 09:21:51 crc kubenswrapper[4696]: E0321 09:21:51.601499 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577" containerName="extract-utilities" Mar 21 09:21:51 crc kubenswrapper[4696]: I0321 09:21:51.601513 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577" containerName="extract-utilities" Mar 21 09:21:51 crc kubenswrapper[4696]: E0321 09:21:51.601524 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577" containerName="registry-server" Mar 21 09:21:51 crc kubenswrapper[4696]: I0321 09:21:51.601529 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577" containerName="registry-server" Mar 21 09:21:51 crc kubenswrapper[4696]: E0321 09:21:51.601537 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7876d0a1-82db-4d12-9b2b-e85160810659" containerName="container-00" Mar 21 09:21:51 crc kubenswrapper[4696]: I0321 09:21:51.601544 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="7876d0a1-82db-4d12-9b2b-e85160810659" containerName="container-00" Mar 21 09:21:51 crc kubenswrapper[4696]: E0321 09:21:51.601554 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577" containerName="extract-content" Mar 21 09:21:51 crc kubenswrapper[4696]: I0321 09:21:51.601559 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577" containerName="extract-content" Mar 21 09:21:51 crc kubenswrapper[4696]: I0321 09:21:51.601733 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3e80b6d-fcfc-42b6-8b50-6fb1d5f51577" containerName="registry-server" Mar 21 09:21:51 crc kubenswrapper[4696]: I0321 09:21:51.601754 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="7876d0a1-82db-4d12-9b2b-e85160810659" containerName="container-00" Mar 21 09:21:51 crc kubenswrapper[4696]: I0321 09:21:51.602438 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mckgv/crc-debug-rq9hw" Mar 21 09:21:51 crc kubenswrapper[4696]: I0321 09:21:51.609903 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mckgv"/"default-dockercfg-fkd4t" Mar 21 09:21:51 crc kubenswrapper[4696]: I0321 09:21:51.670789 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f26a2209-6c48-4339-8e03-c1a5af265d33-host\") pod \"crc-debug-rq9hw\" (UID: \"f26a2209-6c48-4339-8e03-c1a5af265d33\") " pod="openshift-must-gather-mckgv/crc-debug-rq9hw" Mar 21 09:21:51 crc kubenswrapper[4696]: I0321 09:21:51.670984 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6p8f\" (UniqueName: \"kubernetes.io/projected/f26a2209-6c48-4339-8e03-c1a5af265d33-kube-api-access-r6p8f\") pod \"crc-debug-rq9hw\" (UID: \"f26a2209-6c48-4339-8e03-c1a5af265d33\") " pod="openshift-must-gather-mckgv/crc-debug-rq9hw" Mar 21 09:21:51 crc kubenswrapper[4696]: I0321 09:21:51.772975 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f26a2209-6c48-4339-8e03-c1a5af265d33-host\") pod \"crc-debug-rq9hw\" (UID: \"f26a2209-6c48-4339-8e03-c1a5af265d33\") " pod="openshift-must-gather-mckgv/crc-debug-rq9hw" Mar 21 09:21:51 crc kubenswrapper[4696]: I0321 09:21:51.773341 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6p8f\" (UniqueName: \"kubernetes.io/projected/f26a2209-6c48-4339-8e03-c1a5af265d33-kube-api-access-r6p8f\") pod \"crc-debug-rq9hw\" (UID: \"f26a2209-6c48-4339-8e03-c1a5af265d33\") " pod="openshift-must-gather-mckgv/crc-debug-rq9hw" Mar 21 09:21:51 crc kubenswrapper[4696]: I0321 09:21:51.773111 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f26a2209-6c48-4339-8e03-c1a5af265d33-host\") pod \"crc-debug-rq9hw\" (UID: \"f26a2209-6c48-4339-8e03-c1a5af265d33\") " pod="openshift-must-gather-mckgv/crc-debug-rq9hw" Mar 21 09:21:51 crc kubenswrapper[4696]: I0321 09:21:51.793337 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6p8f\" (UniqueName: \"kubernetes.io/projected/f26a2209-6c48-4339-8e03-c1a5af265d33-kube-api-access-r6p8f\") pod \"crc-debug-rq9hw\" (UID: \"f26a2209-6c48-4339-8e03-c1a5af265d33\") " pod="openshift-must-gather-mckgv/crc-debug-rq9hw" Mar 21 09:21:51 crc kubenswrapper[4696]: I0321 09:21:51.979216 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mckgv/crc-debug-rq9hw" Mar 21 09:21:52 crc kubenswrapper[4696]: I0321 09:21:52.120052 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mckgv/crc-debug-rq9hw" event={"ID":"f26a2209-6c48-4339-8e03-c1a5af265d33","Type":"ContainerStarted","Data":"940ac0c6a7f0c7150c67756850c5df2561b9c218d04f5e0e67a95c7fb58546d8"} Mar 21 09:21:52 crc kubenswrapper[4696]: E0321 09:21:52.486586 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf26a2209_6c48_4339_8e03_c1a5af265d33.slice/crio-conmon-32fe9a54955afb311b9b040b66fe6186f25b6657c15ec6a3ac7b1c3c9993956f.scope\": RecentStats: unable to find data in memory cache]" Mar 21 09:21:53 crc kubenswrapper[4696]: I0321 09:21:53.129966 4696 generic.go:334] "Generic (PLEG): container finished" podID="f26a2209-6c48-4339-8e03-c1a5af265d33" containerID="32fe9a54955afb311b9b040b66fe6186f25b6657c15ec6a3ac7b1c3c9993956f" exitCode=0 Mar 21 09:21:53 crc kubenswrapper[4696]: I0321 09:21:53.130072 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mckgv/crc-debug-rq9hw" event={"ID":"f26a2209-6c48-4339-8e03-c1a5af265d33","Type":"ContainerDied","Data":"32fe9a54955afb311b9b040b66fe6186f25b6657c15ec6a3ac7b1c3c9993956f"} Mar 21 09:21:53 crc kubenswrapper[4696]: I0321 09:21:53.967648 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k94px"] Mar 21 09:21:53 crc kubenswrapper[4696]: I0321 09:21:53.969740 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k94px" Mar 21 09:21:54 crc kubenswrapper[4696]: I0321 09:21:54.009555 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k94px"] Mar 21 09:21:54 crc kubenswrapper[4696]: I0321 09:21:54.023034 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv6fd\" (UniqueName: \"kubernetes.io/projected/681596e5-4b49-43ec-8939-f94736fc1047-kube-api-access-qv6fd\") pod \"redhat-marketplace-k94px\" (UID: \"681596e5-4b49-43ec-8939-f94736fc1047\") " pod="openshift-marketplace/redhat-marketplace-k94px" Mar 21 09:21:54 crc kubenswrapper[4696]: I0321 09:21:54.023076 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/681596e5-4b49-43ec-8939-f94736fc1047-catalog-content\") pod \"redhat-marketplace-k94px\" (UID: \"681596e5-4b49-43ec-8939-f94736fc1047\") " pod="openshift-marketplace/redhat-marketplace-k94px" Mar 21 09:21:54 crc kubenswrapper[4696]: I0321 09:21:54.023183 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/681596e5-4b49-43ec-8939-f94736fc1047-utilities\") pod \"redhat-marketplace-k94px\" (UID: \"681596e5-4b49-43ec-8939-f94736fc1047\") " pod="openshift-marketplace/redhat-marketplace-k94px" Mar 21 09:21:54 crc kubenswrapper[4696]: I0321 09:21:54.124763 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv6fd\" (UniqueName: \"kubernetes.io/projected/681596e5-4b49-43ec-8939-f94736fc1047-kube-api-access-qv6fd\") pod \"redhat-marketplace-k94px\" (UID: \"681596e5-4b49-43ec-8939-f94736fc1047\") " pod="openshift-marketplace/redhat-marketplace-k94px" Mar 21 09:21:54 crc kubenswrapper[4696]: I0321 09:21:54.124913 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/681596e5-4b49-43ec-8939-f94736fc1047-catalog-content\") pod \"redhat-marketplace-k94px\" (UID: \"681596e5-4b49-43ec-8939-f94736fc1047\") " pod="openshift-marketplace/redhat-marketplace-k94px" Mar 21 09:21:54 crc kubenswrapper[4696]: I0321 09:21:54.125085 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/681596e5-4b49-43ec-8939-f94736fc1047-utilities\") pod \"redhat-marketplace-k94px\" (UID: \"681596e5-4b49-43ec-8939-f94736fc1047\") " pod="openshift-marketplace/redhat-marketplace-k94px" Mar 21 09:21:54 crc kubenswrapper[4696]: I0321 09:21:54.125682 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/681596e5-4b49-43ec-8939-f94736fc1047-utilities\") pod \"redhat-marketplace-k94px\" (UID: \"681596e5-4b49-43ec-8939-f94736fc1047\") " pod="openshift-marketplace/redhat-marketplace-k94px" Mar 21 09:21:54 crc kubenswrapper[4696]: I0321 09:21:54.126282 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/681596e5-4b49-43ec-8939-f94736fc1047-catalog-content\") pod \"redhat-marketplace-k94px\" (UID: \"681596e5-4b49-43ec-8939-f94736fc1047\") " pod="openshift-marketplace/redhat-marketplace-k94px" Mar 21 09:21:54 crc kubenswrapper[4696]: I0321 09:21:54.165215 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv6fd\" (UniqueName: \"kubernetes.io/projected/681596e5-4b49-43ec-8939-f94736fc1047-kube-api-access-qv6fd\") pod \"redhat-marketplace-k94px\" (UID: \"681596e5-4b49-43ec-8939-f94736fc1047\") " pod="openshift-marketplace/redhat-marketplace-k94px" Mar 21 09:21:54 crc kubenswrapper[4696]: I0321 09:21:54.261551 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mckgv/crc-debug-rq9hw" Mar 21 09:21:54 crc kubenswrapper[4696]: I0321 09:21:54.312032 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mckgv/crc-debug-rq9hw"] Mar 21 09:21:54 crc kubenswrapper[4696]: I0321 09:21:54.321007 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mckgv/crc-debug-rq9hw"] Mar 21 09:21:54 crc kubenswrapper[4696]: I0321 09:21:54.325672 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k94px" Mar 21 09:21:54 crc kubenswrapper[4696]: I0321 09:21:54.328085 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6p8f\" (UniqueName: \"kubernetes.io/projected/f26a2209-6c48-4339-8e03-c1a5af265d33-kube-api-access-r6p8f\") pod \"f26a2209-6c48-4339-8e03-c1a5af265d33\" (UID: \"f26a2209-6c48-4339-8e03-c1a5af265d33\") " Mar 21 09:21:54 crc kubenswrapper[4696]: I0321 09:21:54.328232 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f26a2209-6c48-4339-8e03-c1a5af265d33-host\") pod \"f26a2209-6c48-4339-8e03-c1a5af265d33\" (UID: \"f26a2209-6c48-4339-8e03-c1a5af265d33\") " Mar 21 09:21:54 crc kubenswrapper[4696]: I0321 09:21:54.328844 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f26a2209-6c48-4339-8e03-c1a5af265d33-host" (OuterVolumeSpecName: "host") pod "f26a2209-6c48-4339-8e03-c1a5af265d33" (UID: "f26a2209-6c48-4339-8e03-c1a5af265d33"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 09:21:54 crc kubenswrapper[4696]: I0321 09:21:54.346115 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f26a2209-6c48-4339-8e03-c1a5af265d33-kube-api-access-r6p8f" (OuterVolumeSpecName: "kube-api-access-r6p8f") pod "f26a2209-6c48-4339-8e03-c1a5af265d33" (UID: "f26a2209-6c48-4339-8e03-c1a5af265d33"). InnerVolumeSpecName "kube-api-access-r6p8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:21:54 crc kubenswrapper[4696]: I0321 09:21:54.431200 4696 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f26a2209-6c48-4339-8e03-c1a5af265d33-host\") on node \"crc\" DevicePath \"\"" Mar 21 09:21:54 crc kubenswrapper[4696]: I0321 09:21:54.431465 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6p8f\" (UniqueName: \"kubernetes.io/projected/f26a2209-6c48-4339-8e03-c1a5af265d33-kube-api-access-r6p8f\") on node \"crc\" DevicePath \"\"" Mar 21 09:21:54 crc kubenswrapper[4696]: I0321 09:21:54.571137 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f26a2209-6c48-4339-8e03-c1a5af265d33" path="/var/lib/kubelet/pods/f26a2209-6c48-4339-8e03-c1a5af265d33/volumes" Mar 21 09:21:55 crc kubenswrapper[4696]: I0321 09:21:55.086746 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k94px"] Mar 21 09:21:55 crc kubenswrapper[4696]: W0321 09:21:55.097173 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod681596e5_4b49_43ec_8939_f94736fc1047.slice/crio-af882368b65d6695fc18d5b5cc09e2d826404a4adcbc678222b9c0d8739c600a WatchSource:0}: Error finding container af882368b65d6695fc18d5b5cc09e2d826404a4adcbc678222b9c0d8739c600a: Status 404 returned error can't find the container with id af882368b65d6695fc18d5b5cc09e2d826404a4adcbc678222b9c0d8739c600a Mar 21 09:21:55 crc kubenswrapper[4696]: I0321 09:21:55.158038 4696 scope.go:117] "RemoveContainer" containerID="32fe9a54955afb311b9b040b66fe6186f25b6657c15ec6a3ac7b1c3c9993956f" Mar 21 09:21:55 crc kubenswrapper[4696]: I0321 09:21:55.158162 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mckgv/crc-debug-rq9hw" Mar 21 09:21:55 crc kubenswrapper[4696]: I0321 09:21:55.167271 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k94px" event={"ID":"681596e5-4b49-43ec-8939-f94736fc1047","Type":"ContainerStarted","Data":"af882368b65d6695fc18d5b5cc09e2d826404a4adcbc678222b9c0d8739c600a"} Mar 21 09:21:55 crc kubenswrapper[4696]: I0321 09:21:55.686698 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mckgv/crc-debug-bxrtn"] Mar 21 09:21:55 crc kubenswrapper[4696]: E0321 09:21:55.687557 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f26a2209-6c48-4339-8e03-c1a5af265d33" containerName="container-00" Mar 21 09:21:55 crc kubenswrapper[4696]: I0321 09:21:55.687578 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f26a2209-6c48-4339-8e03-c1a5af265d33" containerName="container-00" Mar 21 09:21:55 crc kubenswrapper[4696]: I0321 09:21:55.687798 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f26a2209-6c48-4339-8e03-c1a5af265d33" containerName="container-00" Mar 21 09:21:55 crc kubenswrapper[4696]: I0321 09:21:55.688712 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mckgv/crc-debug-bxrtn" Mar 21 09:21:55 crc kubenswrapper[4696]: I0321 09:21:55.691217 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mckgv"/"default-dockercfg-fkd4t" Mar 21 09:21:55 crc kubenswrapper[4696]: I0321 09:21:55.759711 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zh8v\" (UniqueName: \"kubernetes.io/projected/88ffa19a-666f-49b5-a425-4f1204e97f42-kube-api-access-4zh8v\") pod \"crc-debug-bxrtn\" (UID: \"88ffa19a-666f-49b5-a425-4f1204e97f42\") " pod="openshift-must-gather-mckgv/crc-debug-bxrtn" Mar 21 09:21:55 crc kubenswrapper[4696]: I0321 09:21:55.760050 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/88ffa19a-666f-49b5-a425-4f1204e97f42-host\") pod \"crc-debug-bxrtn\" (UID: \"88ffa19a-666f-49b5-a425-4f1204e97f42\") " pod="openshift-must-gather-mckgv/crc-debug-bxrtn" Mar 21 09:21:55 crc kubenswrapper[4696]: I0321 09:21:55.875431 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/88ffa19a-666f-49b5-a425-4f1204e97f42-host\") pod \"crc-debug-bxrtn\" (UID: \"88ffa19a-666f-49b5-a425-4f1204e97f42\") " pod="openshift-must-gather-mckgv/crc-debug-bxrtn" Mar 21 09:21:55 crc kubenswrapper[4696]: I0321 09:21:55.876444 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/88ffa19a-666f-49b5-a425-4f1204e97f42-host\") pod \"crc-debug-bxrtn\" (UID: \"88ffa19a-666f-49b5-a425-4f1204e97f42\") " pod="openshift-must-gather-mckgv/crc-debug-bxrtn" Mar 21 09:21:55 crc kubenswrapper[4696]: I0321 09:21:55.911164 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zh8v\" (UniqueName: \"kubernetes.io/projected/88ffa19a-666f-49b5-a425-4f1204e97f42-kube-api-access-4zh8v\") pod \"crc-debug-bxrtn\" (UID: \"88ffa19a-666f-49b5-a425-4f1204e97f42\") " pod="openshift-must-gather-mckgv/crc-debug-bxrtn" Mar 21 09:21:55 crc kubenswrapper[4696]: I0321 09:21:55.947389 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zh8v\" (UniqueName: \"kubernetes.io/projected/88ffa19a-666f-49b5-a425-4f1204e97f42-kube-api-access-4zh8v\") pod \"crc-debug-bxrtn\" (UID: \"88ffa19a-666f-49b5-a425-4f1204e97f42\") " pod="openshift-must-gather-mckgv/crc-debug-bxrtn" Mar 21 09:21:56 crc kubenswrapper[4696]: I0321 09:21:56.005421 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mckgv/crc-debug-bxrtn" Mar 21 09:21:56 crc kubenswrapper[4696]: W0321 09:21:56.037801 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88ffa19a_666f_49b5_a425_4f1204e97f42.slice/crio-987ad796f8386c0f0252452413283d563f25b07606205617572b9829e7f3a32e WatchSource:0}: Error finding container 987ad796f8386c0f0252452413283d563f25b07606205617572b9829e7f3a32e: Status 404 returned error can't find the container with id 987ad796f8386c0f0252452413283d563f25b07606205617572b9829e7f3a32e Mar 21 09:21:56 crc kubenswrapper[4696]: I0321 09:21:56.177303 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mckgv/crc-debug-bxrtn" event={"ID":"88ffa19a-666f-49b5-a425-4f1204e97f42","Type":"ContainerStarted","Data":"987ad796f8386c0f0252452413283d563f25b07606205617572b9829e7f3a32e"} Mar 21 09:21:56 crc kubenswrapper[4696]: I0321 09:21:56.179088 4696 generic.go:334] "Generic (PLEG): container finished" podID="681596e5-4b49-43ec-8939-f94736fc1047" containerID="65d95bd71fbb1cdc4d771c68f533f67b66cb280a56c06e9bf68f62f0e37e9059" exitCode=0 Mar 21 09:21:56 crc kubenswrapper[4696]: I0321 09:21:56.179143 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k94px" event={"ID":"681596e5-4b49-43ec-8939-f94736fc1047","Type":"ContainerDied","Data":"65d95bd71fbb1cdc4d771c68f533f67b66cb280a56c06e9bf68f62f0e37e9059"} Mar 21 09:21:57 crc kubenswrapper[4696]: I0321 09:21:57.195371 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k94px" event={"ID":"681596e5-4b49-43ec-8939-f94736fc1047","Type":"ContainerStarted","Data":"748977f7331d47a4481a3cf07990238e6a118c82e5c4c4ee6e135c612601e6d6"} Mar 21 09:21:57 crc kubenswrapper[4696]: I0321 09:21:57.198184 4696 generic.go:334] "Generic (PLEG): container finished" podID="88ffa19a-666f-49b5-a425-4f1204e97f42" containerID="1d09124d603523ff547a6005dab537e2b746ed60acbee43ebfe1451124c35721" exitCode=0 Mar 21 09:21:57 crc kubenswrapper[4696]: I0321 09:21:57.198277 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mckgv/crc-debug-bxrtn" event={"ID":"88ffa19a-666f-49b5-a425-4f1204e97f42","Type":"ContainerDied","Data":"1d09124d603523ff547a6005dab537e2b746ed60acbee43ebfe1451124c35721"} Mar 21 09:21:57 crc kubenswrapper[4696]: I0321 09:21:57.263231 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mckgv/crc-debug-bxrtn"] Mar 21 09:21:57 crc kubenswrapper[4696]: I0321 09:21:57.273747 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mckgv/crc-debug-bxrtn"] Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.161883 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xx8kz"] Mar 21 09:21:58 crc kubenswrapper[4696]: E0321 09:21:58.162312 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88ffa19a-666f-49b5-a425-4f1204e97f42" containerName="container-00" Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.162328 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="88ffa19a-666f-49b5-a425-4f1204e97f42" containerName="container-00" Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.162531 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="88ffa19a-666f-49b5-a425-4f1204e97f42" containerName="container-00" Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.164273 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xx8kz" Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.180173 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xx8kz"] Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.217324 4696 generic.go:334] "Generic (PLEG): container finished" podID="681596e5-4b49-43ec-8939-f94736fc1047" containerID="748977f7331d47a4481a3cf07990238e6a118c82e5c4c4ee6e135c612601e6d6" exitCode=0 Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.217403 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k94px" event={"ID":"681596e5-4b49-43ec-8939-f94736fc1047","Type":"ContainerDied","Data":"748977f7331d47a4481a3cf07990238e6a118c82e5c4c4ee6e135c612601e6d6"} Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.267886 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a4f80d-c51e-44d3-b823-9f76f1753cbf-catalog-content\") pod \"certified-operators-xx8kz\" (UID: \"d4a4f80d-c51e-44d3-b823-9f76f1753cbf\") " pod="openshift-marketplace/certified-operators-xx8kz" Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.267978 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a4f80d-c51e-44d3-b823-9f76f1753cbf-utilities\") pod \"certified-operators-xx8kz\" (UID: \"d4a4f80d-c51e-44d3-b823-9f76f1753cbf\") " pod="openshift-marketplace/certified-operators-xx8kz" Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.268041 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7knvw\" (UniqueName: \"kubernetes.io/projected/d4a4f80d-c51e-44d3-b823-9f76f1753cbf-kube-api-access-7knvw\") pod \"certified-operators-xx8kz\" (UID: \"d4a4f80d-c51e-44d3-b823-9f76f1753cbf\") " pod="openshift-marketplace/certified-operators-xx8kz" Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.339756 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mckgv/crc-debug-bxrtn" Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.369021 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/88ffa19a-666f-49b5-a425-4f1204e97f42-host\") pod \"88ffa19a-666f-49b5-a425-4f1204e97f42\" (UID: \"88ffa19a-666f-49b5-a425-4f1204e97f42\") " Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.369129 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88ffa19a-666f-49b5-a425-4f1204e97f42-host" (OuterVolumeSpecName: "host") pod "88ffa19a-666f-49b5-a425-4f1204e97f42" (UID: "88ffa19a-666f-49b5-a425-4f1204e97f42"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.369664 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zh8v\" (UniqueName: \"kubernetes.io/projected/88ffa19a-666f-49b5-a425-4f1204e97f42-kube-api-access-4zh8v\") pod \"88ffa19a-666f-49b5-a425-4f1204e97f42\" (UID: \"88ffa19a-666f-49b5-a425-4f1204e97f42\") " Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.370385 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7knvw\" (UniqueName: \"kubernetes.io/projected/d4a4f80d-c51e-44d3-b823-9f76f1753cbf-kube-api-access-7knvw\") pod \"certified-operators-xx8kz\" (UID: \"d4a4f80d-c51e-44d3-b823-9f76f1753cbf\") " pod="openshift-marketplace/certified-operators-xx8kz" Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.370716 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a4f80d-c51e-44d3-b823-9f76f1753cbf-catalog-content\") pod \"certified-operators-xx8kz\" (UID: \"d4a4f80d-c51e-44d3-b823-9f76f1753cbf\") " pod="openshift-marketplace/certified-operators-xx8kz" Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.370893 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a4f80d-c51e-44d3-b823-9f76f1753cbf-utilities\") pod \"certified-operators-xx8kz\" (UID: \"d4a4f80d-c51e-44d3-b823-9f76f1753cbf\") " pod="openshift-marketplace/certified-operators-xx8kz" Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.370990 4696 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/88ffa19a-666f-49b5-a425-4f1204e97f42-host\") on node \"crc\" DevicePath \"\"" Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.371206 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a4f80d-c51e-44d3-b823-9f76f1753cbf-catalog-content\") pod \"certified-operators-xx8kz\" (UID: \"d4a4f80d-c51e-44d3-b823-9f76f1753cbf\") " pod="openshift-marketplace/certified-operators-xx8kz" Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.371260 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a4f80d-c51e-44d3-b823-9f76f1753cbf-utilities\") pod \"certified-operators-xx8kz\" (UID: \"d4a4f80d-c51e-44d3-b823-9f76f1753cbf\") " pod="openshift-marketplace/certified-operators-xx8kz" Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.387185 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88ffa19a-666f-49b5-a425-4f1204e97f42-kube-api-access-4zh8v" (OuterVolumeSpecName: "kube-api-access-4zh8v") pod "88ffa19a-666f-49b5-a425-4f1204e97f42" (UID: "88ffa19a-666f-49b5-a425-4f1204e97f42"). InnerVolumeSpecName "kube-api-access-4zh8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.393580 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7knvw\" (UniqueName: \"kubernetes.io/projected/d4a4f80d-c51e-44d3-b823-9f76f1753cbf-kube-api-access-7knvw\") pod \"certified-operators-xx8kz\" (UID: \"d4a4f80d-c51e-44d3-b823-9f76f1753cbf\") " pod="openshift-marketplace/certified-operators-xx8kz" Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.473056 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zh8v\" (UniqueName: \"kubernetes.io/projected/88ffa19a-666f-49b5-a425-4f1204e97f42-kube-api-access-4zh8v\") on node \"crc\" DevicePath \"\"" Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.481224 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xx8kz" Mar 21 09:21:58 crc kubenswrapper[4696]: I0321 09:21:58.551986 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88ffa19a-666f-49b5-a425-4f1204e97f42" path="/var/lib/kubelet/pods/88ffa19a-666f-49b5-a425-4f1204e97f42/volumes" Mar 21 09:21:59 crc kubenswrapper[4696]: I0321 09:21:59.230006 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k94px" event={"ID":"681596e5-4b49-43ec-8939-f94736fc1047","Type":"ContainerStarted","Data":"6d0398efe80af3b474ce215f07470b38453ec36f72c6360df74bf4cf482320d1"} Mar 21 09:21:59 crc kubenswrapper[4696]: I0321 09:21:59.273935 4696 scope.go:117] "RemoveContainer" containerID="1d09124d603523ff547a6005dab537e2b746ed60acbee43ebfe1451124c35721" Mar 21 09:21:59 crc kubenswrapper[4696]: I0321 09:21:59.274435 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mckgv/crc-debug-bxrtn" Mar 21 09:21:59 crc kubenswrapper[4696]: I0321 09:21:59.320436 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k94px" podStartSLOduration=3.873725142 podStartE2EDuration="6.320414683s" podCreationTimestamp="2026-03-21 09:21:53 +0000 UTC" firstStartedPulling="2026-03-21 09:21:56.182007455 +0000 UTC m=+3250.302888168" lastFinishedPulling="2026-03-21 09:21:58.628696996 +0000 UTC m=+3252.749577709" observedRunningTime="2026-03-21 09:21:59.289062785 +0000 UTC m=+3253.409943508" watchObservedRunningTime="2026-03-21 09:21:59.320414683 +0000 UTC m=+3253.441295396" Mar 21 09:21:59 crc kubenswrapper[4696]: I0321 09:21:59.352808 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xx8kz"] Mar 21 09:21:59 crc kubenswrapper[4696]: W0321 09:21:59.368301 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4a4f80d_c51e_44d3_b823_9f76f1753cbf.slice/crio-e5df4cedadd1841f6e651c67fc432b131e2096845f71e84806fba5983e74c6f6 WatchSource:0}: Error finding container e5df4cedadd1841f6e651c67fc432b131e2096845f71e84806fba5983e74c6f6: Status 404 returned error can't find the container with id e5df4cedadd1841f6e651c67fc432b131e2096845f71e84806fba5983e74c6f6 Mar 21 09:21:59 crc kubenswrapper[4696]: I0321 09:21:59.534589 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:21:59 crc kubenswrapper[4696]: E0321 09:21:59.535099 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:22:00 crc kubenswrapper[4696]: I0321 09:22:00.149387 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568082-s2tkt"] Mar 21 09:22:00 crc kubenswrapper[4696]: I0321 09:22:00.151294 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568082-s2tkt" Mar 21 09:22:00 crc kubenswrapper[4696]: I0321 09:22:00.155099 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:22:00 crc kubenswrapper[4696]: I0321 09:22:00.155411 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:22:00 crc kubenswrapper[4696]: I0321 09:22:00.155526 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:22:00 crc kubenswrapper[4696]: I0321 09:22:00.162752 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568082-s2tkt"] Mar 21 09:22:00 crc kubenswrapper[4696]: I0321 09:22:00.228477 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frv62\" (UniqueName: \"kubernetes.io/projected/60fffaff-5238-4b84-882b-16a0b4c542f6-kube-api-access-frv62\") pod \"auto-csr-approver-29568082-s2tkt\" (UID: \"60fffaff-5238-4b84-882b-16a0b4c542f6\") " pod="openshift-infra/auto-csr-approver-29568082-s2tkt" Mar 21 09:22:00 crc kubenswrapper[4696]: I0321 09:22:00.289658 4696 generic.go:334] "Generic (PLEG): container finished" podID="d4a4f80d-c51e-44d3-b823-9f76f1753cbf" containerID="fb26dfbdcfad89ef52026e93b4ede4c1d68a40d9dd2144659c3224197cf8521a" exitCode=0 Mar 21 09:22:00 crc kubenswrapper[4696]: I0321 09:22:00.292409 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xx8kz" event={"ID":"d4a4f80d-c51e-44d3-b823-9f76f1753cbf","Type":"ContainerDied","Data":"fb26dfbdcfad89ef52026e93b4ede4c1d68a40d9dd2144659c3224197cf8521a"} Mar 21 09:22:00 crc kubenswrapper[4696]: I0321 09:22:00.293150 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xx8kz" event={"ID":"d4a4f80d-c51e-44d3-b823-9f76f1753cbf","Type":"ContainerStarted","Data":"e5df4cedadd1841f6e651c67fc432b131e2096845f71e84806fba5983e74c6f6"} Mar 21 09:22:00 crc kubenswrapper[4696]: I0321 09:22:00.330908 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frv62\" (UniqueName: \"kubernetes.io/projected/60fffaff-5238-4b84-882b-16a0b4c542f6-kube-api-access-frv62\") pod \"auto-csr-approver-29568082-s2tkt\" (UID: \"60fffaff-5238-4b84-882b-16a0b4c542f6\") " pod="openshift-infra/auto-csr-approver-29568082-s2tkt" Mar 21 09:22:00 crc kubenswrapper[4696]: I0321 09:22:00.354003 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frv62\" (UniqueName: \"kubernetes.io/projected/60fffaff-5238-4b84-882b-16a0b4c542f6-kube-api-access-frv62\") pod \"auto-csr-approver-29568082-s2tkt\" (UID: \"60fffaff-5238-4b84-882b-16a0b4c542f6\") " pod="openshift-infra/auto-csr-approver-29568082-s2tkt" Mar 21 09:22:00 crc kubenswrapper[4696]: I0321 09:22:00.471530 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568082-s2tkt" Mar 21 09:22:01 crc kubenswrapper[4696]: I0321 09:22:01.207673 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568082-s2tkt"] Mar 21 09:22:01 crc kubenswrapper[4696]: I0321 09:22:01.316116 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xx8kz" event={"ID":"d4a4f80d-c51e-44d3-b823-9f76f1753cbf","Type":"ContainerStarted","Data":"28e94985e5bb7aa4ee6e6f688a23b26401b54d2aef6604c6389b7522300d8990"} Mar 21 09:22:01 crc kubenswrapper[4696]: I0321 09:22:01.323738 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568082-s2tkt" event={"ID":"60fffaff-5238-4b84-882b-16a0b4c542f6","Type":"ContainerStarted","Data":"a1c37bb7afa95e4681eab8c6e7084bb1e4d7df5c938a23798956acf68d489f67"} Mar 21 09:22:03 crc kubenswrapper[4696]: I0321 09:22:03.345543 4696 generic.go:334] "Generic (PLEG): container finished" podID="d4a4f80d-c51e-44d3-b823-9f76f1753cbf" containerID="28e94985e5bb7aa4ee6e6f688a23b26401b54d2aef6604c6389b7522300d8990" exitCode=0 Mar 21 09:22:03 crc kubenswrapper[4696]: I0321 09:22:03.345635 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xx8kz" event={"ID":"d4a4f80d-c51e-44d3-b823-9f76f1753cbf","Type":"ContainerDied","Data":"28e94985e5bb7aa4ee6e6f688a23b26401b54d2aef6604c6389b7522300d8990"} Mar 21 09:22:03 crc kubenswrapper[4696]: I0321 09:22:03.348541 4696 generic.go:334] "Generic (PLEG): container finished" podID="60fffaff-5238-4b84-882b-16a0b4c542f6" containerID="9e183804d43fe19470aed23665c5decfa6392ab09c92190bb8893e5546519f82" exitCode=0 Mar 21 09:22:03 crc kubenswrapper[4696]: I0321 09:22:03.348575 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568082-s2tkt" event={"ID":"60fffaff-5238-4b84-882b-16a0b4c542f6","Type":"ContainerDied","Data":"9e183804d43fe19470aed23665c5decfa6392ab09c92190bb8893e5546519f82"} Mar 21 09:22:04 crc kubenswrapper[4696]: I0321 09:22:04.326694 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k94px" Mar 21 09:22:04 crc kubenswrapper[4696]: I0321 09:22:04.327075 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k94px" Mar 21 09:22:04 crc kubenswrapper[4696]: I0321 09:22:04.368412 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xx8kz" event={"ID":"d4a4f80d-c51e-44d3-b823-9f76f1753cbf","Type":"ContainerStarted","Data":"f1e494aaa0989fc8b14d2f1c44bc2f176de2cc4e01eaa8d012e43d11b0548ae1"} Mar 21 09:22:04 crc kubenswrapper[4696]: I0321 09:22:04.387099 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k94px" Mar 21 09:22:04 crc kubenswrapper[4696]: I0321 09:22:04.424194 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xx8kz" podStartSLOduration=2.948198534 podStartE2EDuration="6.424175544s" podCreationTimestamp="2026-03-21 09:21:58 +0000 UTC" firstStartedPulling="2026-03-21 09:22:00.293689147 +0000 UTC m=+3254.414569900" lastFinishedPulling="2026-03-21 09:22:03.769666197 +0000 UTC m=+3257.890546910" observedRunningTime="2026-03-21 09:22:04.390617894 +0000 UTC m=+3258.511498607" watchObservedRunningTime="2026-03-21 09:22:04.424175544 +0000 UTC m=+3258.545056257" Mar 21 09:22:04 crc kubenswrapper[4696]: I0321 09:22:04.472432 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k94px" Mar 21 09:22:05 crc kubenswrapper[4696]: I0321 09:22:05.609737 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568082-s2tkt" Mar 21 09:22:05 crc kubenswrapper[4696]: I0321 09:22:05.661217 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frv62\" (UniqueName: \"kubernetes.io/projected/60fffaff-5238-4b84-882b-16a0b4c542f6-kube-api-access-frv62\") pod \"60fffaff-5238-4b84-882b-16a0b4c542f6\" (UID: \"60fffaff-5238-4b84-882b-16a0b4c542f6\") " Mar 21 09:22:05 crc kubenswrapper[4696]: I0321 09:22:05.667019 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60fffaff-5238-4b84-882b-16a0b4c542f6-kube-api-access-frv62" (OuterVolumeSpecName: "kube-api-access-frv62") pod "60fffaff-5238-4b84-882b-16a0b4c542f6" (UID: "60fffaff-5238-4b84-882b-16a0b4c542f6"). InnerVolumeSpecName "kube-api-access-frv62". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:22:05 crc kubenswrapper[4696]: I0321 09:22:05.763765 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frv62\" (UniqueName: \"kubernetes.io/projected/60fffaff-5238-4b84-882b-16a0b4c542f6-kube-api-access-frv62\") on node \"crc\" DevicePath \"\"" Mar 21 09:22:06 crc kubenswrapper[4696]: I0321 09:22:06.396748 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568082-s2tkt" event={"ID":"60fffaff-5238-4b84-882b-16a0b4c542f6","Type":"ContainerDied","Data":"a1c37bb7afa95e4681eab8c6e7084bb1e4d7df5c938a23798956acf68d489f67"} Mar 21 09:22:06 crc kubenswrapper[4696]: I0321 09:22:06.396996 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1c37bb7afa95e4681eab8c6e7084bb1e4d7df5c938a23798956acf68d489f67" Mar 21 09:22:06 crc kubenswrapper[4696]: I0321 09:22:06.396855 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568082-s2tkt" Mar 21 09:22:06 crc kubenswrapper[4696]: I0321 09:22:06.694986 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568076-pchmg"] Mar 21 09:22:06 crc kubenswrapper[4696]: I0321 09:22:06.742174 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568076-pchmg"] Mar 21 09:22:07 crc kubenswrapper[4696]: I0321 09:22:07.354843 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k94px"] Mar 21 09:22:07 crc kubenswrapper[4696]: I0321 09:22:07.355057 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-k94px" podUID="681596e5-4b49-43ec-8939-f94736fc1047" containerName="registry-server" containerID="cri-o://6d0398efe80af3b474ce215f07470b38453ec36f72c6360df74bf4cf482320d1" gracePeriod=2 Mar 21 09:22:08 crc kubenswrapper[4696]: I0321 09:22:08.415785 4696 generic.go:334] "Generic (PLEG): container finished" podID="681596e5-4b49-43ec-8939-f94736fc1047" containerID="6d0398efe80af3b474ce215f07470b38453ec36f72c6360df74bf4cf482320d1" exitCode=0 Mar 21 09:22:08 crc kubenswrapper[4696]: I0321 09:22:08.416034 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k94px" event={"ID":"681596e5-4b49-43ec-8939-f94736fc1047","Type":"ContainerDied","Data":"6d0398efe80af3b474ce215f07470b38453ec36f72c6360df74bf4cf482320d1"} Mar 21 09:22:08 crc kubenswrapper[4696]: I0321 09:22:08.481731 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xx8kz" Mar 21 09:22:08 crc kubenswrapper[4696]: I0321 09:22:08.481805 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xx8kz" Mar 21 09:22:08 crc kubenswrapper[4696]: I0321 09:22:08.545808 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20f13cbd-092d-4a84-97b5-3bb0b48b8e4d" path="/var/lib/kubelet/pods/20f13cbd-092d-4a84-97b5-3bb0b48b8e4d/volumes" Mar 21 09:22:08 crc kubenswrapper[4696]: I0321 09:22:08.691090 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k94px" Mar 21 09:22:08 crc kubenswrapper[4696]: I0321 09:22:08.848657 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qv6fd\" (UniqueName: \"kubernetes.io/projected/681596e5-4b49-43ec-8939-f94736fc1047-kube-api-access-qv6fd\") pod \"681596e5-4b49-43ec-8939-f94736fc1047\" (UID: \"681596e5-4b49-43ec-8939-f94736fc1047\") " Mar 21 09:22:08 crc kubenswrapper[4696]: I0321 09:22:08.848731 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/681596e5-4b49-43ec-8939-f94736fc1047-catalog-content\") pod \"681596e5-4b49-43ec-8939-f94736fc1047\" (UID: \"681596e5-4b49-43ec-8939-f94736fc1047\") " Mar 21 09:22:08 crc kubenswrapper[4696]: I0321 09:22:08.848769 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/681596e5-4b49-43ec-8939-f94736fc1047-utilities\") pod \"681596e5-4b49-43ec-8939-f94736fc1047\" (UID: \"681596e5-4b49-43ec-8939-f94736fc1047\") " Mar 21 09:22:08 crc kubenswrapper[4696]: I0321 09:22:08.851751 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/681596e5-4b49-43ec-8939-f94736fc1047-utilities" (OuterVolumeSpecName: "utilities") pod "681596e5-4b49-43ec-8939-f94736fc1047" (UID: "681596e5-4b49-43ec-8939-f94736fc1047"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:22:08 crc kubenswrapper[4696]: I0321 09:22:08.861638 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/681596e5-4b49-43ec-8939-f94736fc1047-kube-api-access-qv6fd" (OuterVolumeSpecName: "kube-api-access-qv6fd") pod "681596e5-4b49-43ec-8939-f94736fc1047" (UID: "681596e5-4b49-43ec-8939-f94736fc1047"). InnerVolumeSpecName "kube-api-access-qv6fd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:22:08 crc kubenswrapper[4696]: I0321 09:22:08.876859 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/681596e5-4b49-43ec-8939-f94736fc1047-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "681596e5-4b49-43ec-8939-f94736fc1047" (UID: "681596e5-4b49-43ec-8939-f94736fc1047"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:22:08 crc kubenswrapper[4696]: I0321 09:22:08.951424 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qv6fd\" (UniqueName: \"kubernetes.io/projected/681596e5-4b49-43ec-8939-f94736fc1047-kube-api-access-qv6fd\") on node \"crc\" DevicePath \"\"" Mar 21 09:22:08 crc kubenswrapper[4696]: I0321 09:22:08.951460 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/681596e5-4b49-43ec-8939-f94736fc1047-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 09:22:08 crc kubenswrapper[4696]: I0321 09:22:08.951469 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/681596e5-4b49-43ec-8939-f94736fc1047-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 09:22:09 crc kubenswrapper[4696]: I0321 09:22:09.428122 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k94px" event={"ID":"681596e5-4b49-43ec-8939-f94736fc1047","Type":"ContainerDied","Data":"af882368b65d6695fc18d5b5cc09e2d826404a4adcbc678222b9c0d8739c600a"} Mar 21 09:22:09 crc kubenswrapper[4696]: I0321 09:22:09.428176 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k94px" Mar 21 09:22:09 crc kubenswrapper[4696]: I0321 09:22:09.428182 4696 scope.go:117] "RemoveContainer" containerID="6d0398efe80af3b474ce215f07470b38453ec36f72c6360df74bf4cf482320d1" Mar 21 09:22:09 crc kubenswrapper[4696]: I0321 09:22:09.447028 4696 scope.go:117] "RemoveContainer" containerID="748977f7331d47a4481a3cf07990238e6a118c82e5c4c4ee6e135c612601e6d6" Mar 21 09:22:09 crc kubenswrapper[4696]: I0321 09:22:09.465439 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k94px"] Mar 21 09:22:09 crc kubenswrapper[4696]: I0321 09:22:09.471068 4696 scope.go:117] "RemoveContainer" containerID="65d95bd71fbb1cdc4d771c68f533f67b66cb280a56c06e9bf68f62f0e37e9059" Mar 21 09:22:09 crc kubenswrapper[4696]: I0321 09:22:09.486086 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-k94px"] Mar 21 09:22:09 crc kubenswrapper[4696]: I0321 09:22:09.531367 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-xx8kz" podUID="d4a4f80d-c51e-44d3-b823-9f76f1753cbf" containerName="registry-server" probeResult="failure" output=< Mar 21 09:22:09 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Mar 21 09:22:09 crc kubenswrapper[4696]: > Mar 21 09:22:10 crc kubenswrapper[4696]: I0321 09:22:10.536624 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:22:10 crc kubenswrapper[4696]: E0321 09:22:10.537191 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:22:10 crc kubenswrapper[4696]: I0321 09:22:10.551276 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="681596e5-4b49-43ec-8939-f94736fc1047" path="/var/lib/kubelet/pods/681596e5-4b49-43ec-8939-f94736fc1047/volumes" Mar 21 09:22:18 crc kubenswrapper[4696]: I0321 09:22:18.569686 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xx8kz" Mar 21 09:22:18 crc kubenswrapper[4696]: I0321 09:22:18.637509 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xx8kz" Mar 21 09:22:19 crc kubenswrapper[4696]: I0321 09:22:19.559280 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xx8kz"] Mar 21 09:22:20 crc kubenswrapper[4696]: I0321 09:22:20.525151 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xx8kz" podUID="d4a4f80d-c51e-44d3-b823-9f76f1753cbf" containerName="registry-server" containerID="cri-o://f1e494aaa0989fc8b14d2f1c44bc2f176de2cc4e01eaa8d012e43d11b0548ae1" gracePeriod=2 Mar 21 09:22:21 crc kubenswrapper[4696]: I0321 09:22:21.534888 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:22:21 crc kubenswrapper[4696]: E0321 09:22:21.535537 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:22:21 crc kubenswrapper[4696]: I0321 09:22:21.537113 4696 generic.go:334] "Generic (PLEG): container finished" podID="d4a4f80d-c51e-44d3-b823-9f76f1753cbf" containerID="f1e494aaa0989fc8b14d2f1c44bc2f176de2cc4e01eaa8d012e43d11b0548ae1" exitCode=0 Mar 21 09:22:21 crc kubenswrapper[4696]: I0321 09:22:21.537165 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xx8kz" event={"ID":"d4a4f80d-c51e-44d3-b823-9f76f1753cbf","Type":"ContainerDied","Data":"f1e494aaa0989fc8b14d2f1c44bc2f176de2cc4e01eaa8d012e43d11b0548ae1"} Mar 21 09:22:21 crc kubenswrapper[4696]: I0321 09:22:21.687288 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xx8kz" Mar 21 09:22:21 crc kubenswrapper[4696]: I0321 09:22:21.783187 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a4f80d-c51e-44d3-b823-9f76f1753cbf-catalog-content\") pod \"d4a4f80d-c51e-44d3-b823-9f76f1753cbf\" (UID: \"d4a4f80d-c51e-44d3-b823-9f76f1753cbf\") " Mar 21 09:22:21 crc kubenswrapper[4696]: I0321 09:22:21.783347 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a4f80d-c51e-44d3-b823-9f76f1753cbf-utilities\") pod \"d4a4f80d-c51e-44d3-b823-9f76f1753cbf\" (UID: \"d4a4f80d-c51e-44d3-b823-9f76f1753cbf\") " Mar 21 09:22:21 crc kubenswrapper[4696]: I0321 09:22:21.783428 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7knvw\" (UniqueName: \"kubernetes.io/projected/d4a4f80d-c51e-44d3-b823-9f76f1753cbf-kube-api-access-7knvw\") pod \"d4a4f80d-c51e-44d3-b823-9f76f1753cbf\" (UID: \"d4a4f80d-c51e-44d3-b823-9f76f1753cbf\") " Mar 21 09:22:21 crc kubenswrapper[4696]: I0321 09:22:21.785091 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4a4f80d-c51e-44d3-b823-9f76f1753cbf-utilities" (OuterVolumeSpecName: "utilities") pod "d4a4f80d-c51e-44d3-b823-9f76f1753cbf" (UID: "d4a4f80d-c51e-44d3-b823-9f76f1753cbf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:22:21 crc kubenswrapper[4696]: I0321 09:22:21.792196 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4a4f80d-c51e-44d3-b823-9f76f1753cbf-kube-api-access-7knvw" (OuterVolumeSpecName: "kube-api-access-7knvw") pod "d4a4f80d-c51e-44d3-b823-9f76f1753cbf" (UID: "d4a4f80d-c51e-44d3-b823-9f76f1753cbf"). InnerVolumeSpecName "kube-api-access-7knvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:22:21 crc kubenswrapper[4696]: I0321 09:22:21.886300 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a4f80d-c51e-44d3-b823-9f76f1753cbf-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 09:22:21 crc kubenswrapper[4696]: I0321 09:22:21.886337 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7knvw\" (UniqueName: \"kubernetes.io/projected/d4a4f80d-c51e-44d3-b823-9f76f1753cbf-kube-api-access-7knvw\") on node \"crc\" DevicePath \"\"" Mar 21 09:22:21 crc kubenswrapper[4696]: I0321 09:22:21.886482 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4a4f80d-c51e-44d3-b823-9f76f1753cbf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4a4f80d-c51e-44d3-b823-9f76f1753cbf" (UID: "d4a4f80d-c51e-44d3-b823-9f76f1753cbf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:22:21 crc kubenswrapper[4696]: I0321 09:22:21.987770 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a4f80d-c51e-44d3-b823-9f76f1753cbf-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 09:22:22 crc kubenswrapper[4696]: I0321 09:22:22.548867 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xx8kz" event={"ID":"d4a4f80d-c51e-44d3-b823-9f76f1753cbf","Type":"ContainerDied","Data":"e5df4cedadd1841f6e651c67fc432b131e2096845f71e84806fba5983e74c6f6"} Mar 21 09:22:22 crc kubenswrapper[4696]: I0321 09:22:22.548917 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xx8kz" Mar 21 09:22:22 crc kubenswrapper[4696]: I0321 09:22:22.548921 4696 scope.go:117] "RemoveContainer" containerID="f1e494aaa0989fc8b14d2f1c44bc2f176de2cc4e01eaa8d012e43d11b0548ae1" Mar 21 09:22:22 crc kubenswrapper[4696]: I0321 09:22:22.570561 4696 scope.go:117] "RemoveContainer" containerID="28e94985e5bb7aa4ee6e6f688a23b26401b54d2aef6604c6389b7522300d8990" Mar 21 09:22:22 crc kubenswrapper[4696]: I0321 09:22:22.589122 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xx8kz"] Mar 21 09:22:22 crc kubenswrapper[4696]: I0321 09:22:22.600085 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xx8kz"] Mar 21 09:22:22 crc kubenswrapper[4696]: I0321 09:22:22.601272 4696 scope.go:117] "RemoveContainer" containerID="fb26dfbdcfad89ef52026e93b4ede4c1d68a40d9dd2144659c3224197cf8521a" Mar 21 09:22:24 crc kubenswrapper[4696]: I0321 09:22:24.545885 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4a4f80d-c51e-44d3-b823-9f76f1753cbf" path="/var/lib/kubelet/pods/d4a4f80d-c51e-44d3-b823-9f76f1753cbf/volumes" Mar 21 09:22:35 crc kubenswrapper[4696]: I0321 09:22:35.534886 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:22:35 crc kubenswrapper[4696]: E0321 09:22:35.535599 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:22:45 crc kubenswrapper[4696]: I0321 09:22:45.309671 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_b1808d81-bf6f-4201-972a-507d6b41f33f/init-config-reloader/0.log" Mar 21 09:22:45 crc kubenswrapper[4696]: I0321 09:22:45.567154 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_b1808d81-bf6f-4201-972a-507d6b41f33f/init-config-reloader/0.log" Mar 21 09:22:45 crc kubenswrapper[4696]: I0321 09:22:45.616429 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_b1808d81-bf6f-4201-972a-507d6b41f33f/config-reloader/0.log" Mar 21 09:22:45 crc kubenswrapper[4696]: I0321 09:22:45.724897 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_b1808d81-bf6f-4201-972a-507d6b41f33f/alertmanager/0.log" Mar 21 09:22:45 crc kubenswrapper[4696]: I0321 09:22:45.878971 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7dd88fbdb6-nnxhg_bd0e5187-2172-48fc-90e8-4ce8f3f1c143/barbican-api-log/0.log" Mar 21 09:22:45 crc kubenswrapper[4696]: I0321 09:22:45.936079 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7dd88fbdb6-nnxhg_bd0e5187-2172-48fc-90e8-4ce8f3f1c143/barbican-api/0.log" Mar 21 09:22:46 crc kubenswrapper[4696]: I0321 09:22:46.135487 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7c894bbd86-9phxx_adeff71a-2b97-46e9-a945-d69cf06e34d9/barbican-keystone-listener/0.log" Mar 21 09:22:46 crc kubenswrapper[4696]: I0321 09:22:46.291168 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7c894bbd86-9phxx_adeff71a-2b97-46e9-a945-d69cf06e34d9/barbican-keystone-listener-log/0.log" Mar 21 09:22:46 crc kubenswrapper[4696]: I0321 09:22:46.460662 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-756466fd6c-6vdmf_639973de-e4aa-46d4-817a-f9c410238b45/barbican-worker-log/0.log" Mar 21 09:22:46 crc kubenswrapper[4696]: I0321 09:22:46.509831 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-756466fd6c-6vdmf_639973de-e4aa-46d4-817a-f9c410238b45/barbican-worker/0.log" Mar 21 09:22:46 crc kubenswrapper[4696]: I0321 09:22:46.770949 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq_875486e6-3369-4601-bc53-00c2aaf85707/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:22:47 crc kubenswrapper[4696]: I0321 09:22:47.039463 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_72bd0225-bcbe-44ec-be2a-4f38093ea9c9/ceilometer-central-agent/0.log" Mar 21 09:22:47 crc kubenswrapper[4696]: I0321 09:22:47.233119 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_72bd0225-bcbe-44ec-be2a-4f38093ea9c9/proxy-httpd/0.log" Mar 21 09:22:47 crc kubenswrapper[4696]: I0321 09:22:47.321400 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_72bd0225-bcbe-44ec-be2a-4f38093ea9c9/sg-core/0.log" Mar 21 09:22:47 crc kubenswrapper[4696]: I0321 09:22:47.363398 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_72bd0225-bcbe-44ec-be2a-4f38093ea9c9/ceilometer-notification-agent/0.log" Mar 21 09:22:47 crc kubenswrapper[4696]: I0321 09:22:47.534368 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:22:47 crc kubenswrapper[4696]: E0321 09:22:47.535177 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:22:47 crc kubenswrapper[4696]: I0321 09:22:47.619651 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8d228678-e5cf-4da4-9a96-7ed39576c142/cinder-api/0.log" Mar 21 09:22:47 crc kubenswrapper[4696]: I0321 09:22:47.702170 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8d228678-e5cf-4da4-9a96-7ed39576c142/cinder-api-log/0.log" Mar 21 09:22:47 crc kubenswrapper[4696]: I0321 09:22:47.906153 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4fce4b85-2f39-4702-b221-fd5e870e6a2d/cinder-scheduler/0.log" Mar 21 09:22:48 crc kubenswrapper[4696]: I0321 09:22:48.106321 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4fce4b85-2f39-4702-b221-fd5e870e6a2d/probe/0.log" Mar 21 09:22:48 crc kubenswrapper[4696]: I0321 09:22:48.205073 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_16a5bae2-5738-4f3c-a6ca-1e08aa8930a0/cloudkitty-api-log/0.log" Mar 21 09:22:48 crc kubenswrapper[4696]: I0321 09:22:48.261149 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_16a5bae2-5738-4f3c-a6ca-1e08aa8930a0/cloudkitty-api/0.log" Mar 21 09:22:48 crc kubenswrapper[4696]: I0321 09:22:48.499503 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-compactor-0_14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8/loki-compactor/0.log" Mar 21 09:22:48 crc kubenswrapper[4696]: I0321 09:22:48.736733 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7_a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7/loki-distributor/0.log" Mar 21 09:22:48 crc kubenswrapper[4696]: I0321 09:22:48.973556 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-6b884dc4b5-lczl8_d2b0b7e1-1fa0-44ce-bab0-0964460e645b/gateway/0.log" Mar 21 09:22:49 crc kubenswrapper[4696]: I0321 09:22:49.431914 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w_70a97941-5dea-48a7-8dbf-b475b4af0b6e/gateway/0.log" Mar 21 09:22:49 crc kubenswrapper[4696]: I0321 09:22:49.635746 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-index-gateway-0_88865f3d-9bb9-4844-bb96-dc25c11615ec/loki-index-gateway/0.log" Mar 21 09:22:50 crc kubenswrapper[4696]: I0321 09:22:50.027147 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-ingester-0_d6518eba-34ec-4a38-b279-cd8bb11e7e24/loki-ingester/0.log" Mar 21 09:22:50 crc kubenswrapper[4696]: I0321 09:22:50.356450 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-querier-668f98fdd7-tnh77_edf07829-775b-4672-89fe-81c3342300ae/loki-querier/0.log" Mar 21 09:22:50 crc kubenswrapper[4696]: I0321 09:22:50.552906 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-query-frontend-6f54889599-6mn7k_bbf61ae0-6918-44af-be69-ac8a220fcd6b/loki-query-frontend/0.log" Mar 21 09:22:51 crc kubenswrapper[4696]: I0321 09:22:51.025584 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv_719c50b9-02ed-4138-8ce6-624b718e5d00/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:22:51 crc kubenswrapper[4696]: I0321 09:22:51.418160 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-6k46h_7f2cc308-7143-4847-9e6d-a96e9ddd6c51/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:22:51 crc kubenswrapper[4696]: I0321 09:22:51.436593 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-85f64749dc-tqgff_c9fe9d98-b02d-4fdd-918e-c296d52f7efd/init/0.log" Mar 21 09:22:52 crc kubenswrapper[4696]: I0321 09:22:52.206004 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-85f64749dc-tqgff_c9fe9d98-b02d-4fdd-918e-c296d52f7efd/init/0.log" Mar 21 09:22:52 crc kubenswrapper[4696]: I0321 09:22:52.276480 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-zncdp_c756d18c-c925-44bd-911f-0b82d961c66e/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:22:52 crc kubenswrapper[4696]: I0321 09:22:52.393768 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-85f64749dc-tqgff_c9fe9d98-b02d-4fdd-918e-c296d52f7efd/dnsmasq-dns/0.log" Mar 21 09:22:52 crc kubenswrapper[4696]: I0321 09:22:52.704938 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_325e7183-8d8e-4a92-ab49-3dc4a043c21c/glance-log/0.log" Mar 21 09:22:52 crc kubenswrapper[4696]: I0321 09:22:52.733418 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_325e7183-8d8e-4a92-ab49-3dc4a043c21c/glance-httpd/0.log" Mar 21 09:22:53 crc kubenswrapper[4696]: I0321 09:22:53.117977 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7/glance-httpd/0.log" Mar 21 09:22:53 crc kubenswrapper[4696]: I0321 09:22:53.158559 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7/glance-log/0.log" Mar 21 09:22:53 crc kubenswrapper[4696]: I0321 09:22:53.418022 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2_053326a0-54a2-4838-b024-4650e216f804/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:22:53 crc kubenswrapper[4696]: I0321 09:22:53.810136 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-85v2z_3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:22:54 crc kubenswrapper[4696]: I0321 09:22:54.226270 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29568061-9qzjv_93e69ad1-eb81-4a1a-ae60-6feff0891ced/keystone-cron/0.log" Mar 21 09:22:54 crc kubenswrapper[4696]: I0321 09:22:54.307407 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-696f547b5d-r9gcd_31091b53-3bf3-4704-96bc-c5c33eb55218/keystone-api/0.log" Mar 21 09:22:54 crc kubenswrapper[4696]: I0321 09:22:54.338486 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_53c78963-a90b-40f0-b223-0c70917cee6e/kube-state-metrics/0.log" Mar 21 09:22:55 crc kubenswrapper[4696]: I0321 09:22:55.442917 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d_50e8cc4d-e8bf-4f1a-9c84-f40af2392cde/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:22:55 crc kubenswrapper[4696]: I0321 09:22:55.704187 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-69b5b45b6f-7lhn7_1157f1b6-6891-43eb-93fc-e8e66691359a/neutron-api/0.log" Mar 21 09:22:56 crc kubenswrapper[4696]: I0321 09:22:56.012746 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-69b5b45b6f-7lhn7_1157f1b6-6891-43eb-93fc-e8e66691359a/neutron-httpd/0.log" Mar 21 09:22:56 crc kubenswrapper[4696]: I0321 09:22:56.518532 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv_76b63ef6-105a-4d99-a8bc-a8758970ffff/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:22:57 crc kubenswrapper[4696]: I0321 09:22:57.354518 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_348bda58-214e-41e3-bc04-2fdcf0e79142/nova-api-log/0.log" Mar 21 09:22:57 crc kubenswrapper[4696]: I0321 09:22:57.416338 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_348bda58-214e-41e3-bc04-2fdcf0e79142/nova-api-api/0.log" Mar 21 09:22:57 crc kubenswrapper[4696]: I0321 09:22:57.512770 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_7ba86395-973a-4d8b-976f-673403ae58e5/nova-cell0-conductor-conductor/0.log" Mar 21 09:22:57 crc kubenswrapper[4696]: I0321 09:22:57.848148 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267/nova-cell1-conductor-conductor/0.log" Mar 21 09:22:58 crc kubenswrapper[4696]: I0321 09:22:58.002185 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-proc-0_e0413efc-468a-4065-9b3f-6355ec9d0e1e/cloudkitty-proc/0.log" Mar 21 09:22:58 crc kubenswrapper[4696]: I0321 09:22:58.465906 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_f424006b-438e-46cc-8d5f-70d8d0346f40/nova-cell1-novncproxy-novncproxy/0.log" Mar 21 09:22:58 crc kubenswrapper[4696]: I0321 09:22:58.806282 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f812f449-0a51-4414-ac6f-31ecf7648686/nova-metadata-log/0.log" Mar 21 09:22:58 crc kubenswrapper[4696]: I0321 09:22:58.871020 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-dpstl_14ded188-15ec-4d0c-a462-d1160dfb5b47/nova-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:22:59 crc kubenswrapper[4696]: I0321 09:22:59.210674 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f812f449-0a51-4414-ac6f-31ecf7648686/nova-metadata-metadata/0.log" Mar 21 09:22:59 crc kubenswrapper[4696]: I0321 09:22:59.274266 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_57e750dc-d19c-45d4-9905-b0a15a6a642d/mysql-bootstrap/0.log" Mar 21 09:22:59 crc kubenswrapper[4696]: I0321 09:22:59.322896 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_8b1800dd-28c2-4614-be8a-06afc002a85e/nova-scheduler-scheduler/0.log" Mar 21 09:22:59 crc kubenswrapper[4696]: I0321 09:22:59.536156 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:22:59 crc kubenswrapper[4696]: E0321 09:22:59.536465 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:22:59 crc kubenswrapper[4696]: I0321 09:22:59.730556 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_57e750dc-d19c-45d4-9905-b0a15a6a642d/mysql-bootstrap/0.log" Mar 21 09:22:59 crc kubenswrapper[4696]: I0321 09:22:59.790777 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_57e750dc-d19c-45d4-9905-b0a15a6a642d/galera/0.log" Mar 21 09:22:59 crc kubenswrapper[4696]: I0321 09:22:59.836872 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_919a11c7-cf69-4660-8de6-6b1a54b63f26/mysql-bootstrap/0.log" Mar 21 09:23:00 crc kubenswrapper[4696]: I0321 09:23:00.507165 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_919a11c7-cf69-4660-8de6-6b1a54b63f26/galera/0.log" Mar 21 09:23:00 crc kubenswrapper[4696]: I0321 09:23:00.788432 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_383967e9-132f-4355-abbb-5904830137b7/openstackclient/0.log" Mar 21 09:23:00 crc kubenswrapper[4696]: I0321 09:23:00.795103 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_919a11c7-cf69-4660-8de6-6b1a54b63f26/mysql-bootstrap/0.log" Mar 21 09:23:01 crc kubenswrapper[4696]: I0321 09:23:01.217606 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-4vd4t_79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a/ovn-controller/0.log" Mar 21 09:23:01 crc kubenswrapper[4696]: I0321 09:23:01.409858 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-8nzqr_63645afc-9440-40b7-ae49-9203c9dcb3f8/openstack-network-exporter/0.log" Mar 21 09:23:01 crc kubenswrapper[4696]: I0321 09:23:01.687554 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dmbpm_7f0d9af4-d6e3-4168-b83d-5a482fb51d44/ovsdb-server-init/0.log" Mar 21 09:23:01 crc kubenswrapper[4696]: I0321 09:23:01.982931 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dmbpm_7f0d9af4-d6e3-4168-b83d-5a482fb51d44/ovsdb-server-init/0.log" Mar 21 09:23:02 crc kubenswrapper[4696]: I0321 09:23:02.044559 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dmbpm_7f0d9af4-d6e3-4168-b83d-5a482fb51d44/ovsdb-server/0.log" Mar 21 09:23:02 crc kubenswrapper[4696]: I0321 09:23:02.069719 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dmbpm_7f0d9af4-d6e3-4168-b83d-5a482fb51d44/ovs-vswitchd/0.log" Mar 21 09:23:02 crc kubenswrapper[4696]: I0321 09:23:02.399177 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_dc5349db-8b6f-41a8-b3c7-d42920c3bf58/openstack-network-exporter/0.log" Mar 21 09:23:02 crc kubenswrapper[4696]: I0321 09:23:02.539496 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-v5hcz_2c75b9c9-8d94-4541-8858-0c56d8ecb400/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:23:02 crc kubenswrapper[4696]: I0321 09:23:02.684270 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_dc5349db-8b6f-41a8-b3c7-d42920c3bf58/ovn-northd/0.log" Mar 21 09:23:02 crc kubenswrapper[4696]: I0321 09:23:02.803844 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21/openstack-network-exporter/0.log" Mar 21 09:23:02 crc kubenswrapper[4696]: I0321 09:23:02.835272 4696 scope.go:117] "RemoveContainer" containerID="bbd36ab2f3383ee30dc688a02c06ad325d4689455afd4fc7483648a434349e41" Mar 21 09:23:02 crc kubenswrapper[4696]: I0321 09:23:02.919445 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21/ovsdbserver-nb/0.log" Mar 21 09:23:03 crc kubenswrapper[4696]: I0321 09:23:03.194028 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_b9782617-ee6a-4ab5-aa9c-22dc382f654e/openstack-network-exporter/0.log" Mar 21 09:23:03 crc kubenswrapper[4696]: I0321 09:23:03.214500 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_b9782617-ee6a-4ab5-aa9c-22dc382f654e/ovsdbserver-sb/0.log" Mar 21 09:23:03 crc kubenswrapper[4696]: I0321 09:23:03.788808 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-56b7788dc4-f928g_3bd18ca4-e084-4fe1-85e3-7319ddd703ff/placement-api/0.log" Mar 21 09:23:03 crc kubenswrapper[4696]: I0321 09:23:03.968255 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-56b7788dc4-f928g_3bd18ca4-e084-4fe1-85e3-7319ddd703ff/placement-log/0.log" Mar 21 09:23:04 crc kubenswrapper[4696]: I0321 09:23:04.263409 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_c2c63b45-de34-43a6-bb8b-aed4b00922c8/init-config-reloader/0.log" Mar 21 09:23:04 crc kubenswrapper[4696]: I0321 09:23:04.530854 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_c2c63b45-de34-43a6-bb8b-aed4b00922c8/config-reloader/0.log" Mar 21 09:23:04 crc kubenswrapper[4696]: I0321 09:23:04.602031 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_c2c63b45-de34-43a6-bb8b-aed4b00922c8/init-config-reloader/0.log" Mar 21 09:23:04 crc kubenswrapper[4696]: I0321 09:23:04.610273 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_c2c63b45-de34-43a6-bb8b-aed4b00922c8/prometheus/0.log" Mar 21 09:23:04 crc kubenswrapper[4696]: I0321 09:23:04.678686 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_c2c63b45-de34-43a6-bb8b-aed4b00922c8/thanos-sidecar/0.log" Mar 21 09:23:04 crc kubenswrapper[4696]: I0321 09:23:04.851021 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0f40066c-3b43-42a5-9bb2-c8fb94a921db/setup-container/0.log" Mar 21 09:23:05 crc kubenswrapper[4696]: I0321 09:23:05.275096 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0f40066c-3b43-42a5-9bb2-c8fb94a921db/setup-container/0.log" Mar 21 09:23:05 crc kubenswrapper[4696]: I0321 09:23:05.346697 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0f40066c-3b43-42a5-9bb2-c8fb94a921db/rabbitmq/0.log" Mar 21 09:23:05 crc kubenswrapper[4696]: I0321 09:23:05.396719 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_56d6eab2-0e95-402a-afb7-2707ca41b144/setup-container/0.log" Mar 21 09:23:05 crc kubenswrapper[4696]: I0321 09:23:05.848453 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h_e19a30a1-9816-4c71-9803-ed0dbbdad615/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:23:05 crc kubenswrapper[4696]: I0321 09:23:05.861696 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_56d6eab2-0e95-402a-afb7-2707ca41b144/setup-container/0.log" Mar 21 09:23:06 crc kubenswrapper[4696]: I0321 09:23:06.054759 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_56d6eab2-0e95-402a-afb7-2707ca41b144/rabbitmq/0.log" Mar 21 09:23:06 crc kubenswrapper[4696]: I0321 09:23:06.172091 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_dd0d5d6d-801d-4446-a325-f82b2d55bca6/memcached/0.log" Mar 21 09:23:06 crc kubenswrapper[4696]: I0321 09:23:06.292307 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-6cdss_480a9890-311e-44b9-bfac-4c375e7cdf1e/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:23:06 crc kubenswrapper[4696]: I0321 09:23:06.427236 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n_93d9988c-26af-48d1-b43e-7c897cad7235/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:23:06 crc kubenswrapper[4696]: I0321 09:23:06.724776 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-km5wx_5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:23:06 crc kubenswrapper[4696]: I0321 09:23:06.803328 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-ckxp8_4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6/ssh-known-hosts-edpm-deployment/0.log" Mar 21 09:23:06 crc kubenswrapper[4696]: I0321 09:23:06.990054 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7b65745bbc-5b5zz_4abc84ca-2aab-4423-b934-c61f3ef8ea6d/proxy-httpd/0.log" Mar 21 09:23:07 crc kubenswrapper[4696]: I0321 09:23:07.195733 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7b65745bbc-5b5zz_4abc84ca-2aab-4423-b934-c61f3ef8ea6d/proxy-server/0.log" Mar 21 09:23:07 crc kubenswrapper[4696]: I0321 09:23:07.257141 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-jznlt_efd827ca-ae90-4cca-a77d-72fce76c47aa/swift-ring-rebalance/0.log" Mar 21 09:23:07 crc kubenswrapper[4696]: I0321 09:23:07.689489 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/account-auditor/0.log" Mar 21 09:23:07 crc kubenswrapper[4696]: I0321 09:23:07.831539 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/account-reaper/0.log" Mar 21 09:23:07 crc kubenswrapper[4696]: I0321 09:23:07.887842 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/account-server/0.log" Mar 21 09:23:07 crc kubenswrapper[4696]: I0321 09:23:07.944722 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/account-replicator/0.log" Mar 21 09:23:08 crc kubenswrapper[4696]: I0321 09:23:08.051281 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/container-auditor/0.log" Mar 21 09:23:08 crc kubenswrapper[4696]: I0321 09:23:08.061336 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/container-replicator/0.log" Mar 21 09:23:08 crc kubenswrapper[4696]: I0321 09:23:08.103174 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/container-server/0.log" Mar 21 09:23:08 crc kubenswrapper[4696]: I0321 09:23:08.285763 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/object-auditor/0.log" Mar 21 09:23:08 crc kubenswrapper[4696]: I0321 09:23:08.470762 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/object-server/0.log" Mar 21 09:23:08 crc kubenswrapper[4696]: I0321 09:23:08.495464 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/container-updater/0.log" Mar 21 09:23:08 crc kubenswrapper[4696]: I0321 09:23:08.563121 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/object-replicator/0.log" Mar 21 09:23:08 crc kubenswrapper[4696]: I0321 09:23:08.628280 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/object-updater/0.log" Mar 21 09:23:08 crc kubenswrapper[4696]: I0321 09:23:08.668976 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/object-expirer/0.log" Mar 21 09:23:08 crc kubenswrapper[4696]: I0321 09:23:08.794990 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/rsync/0.log" Mar 21 09:23:08 crc kubenswrapper[4696]: I0321 09:23:08.931319 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/swift-recon-cron/0.log" Mar 21 09:23:09 crc kubenswrapper[4696]: I0321 09:23:09.411815 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8_2b3f6677-db77-456b-aff2-9deee1fe26f9/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:23:09 crc kubenswrapper[4696]: I0321 09:23:09.463974 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e/tempest-tests-tempest-tests-runner/0.log" Mar 21 09:23:09 crc kubenswrapper[4696]: I0321 09:23:09.636145 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_e6c3f461-61e5-443c-911d-0b8a1ba72a96/test-operator-logs-container/0.log" Mar 21 09:23:09 crc kubenswrapper[4696]: I0321 09:23:09.710230 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8_a8cdd5da-1ba0-4034-8a15-e3e86352772f/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:23:11 crc kubenswrapper[4696]: I0321 09:23:11.534790 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:23:12 crc kubenswrapper[4696]: I0321 09:23:12.109557 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerStarted","Data":"aa5a03c21bdd6334cc43af15b5fd004e41076b2d7b35f63b8f51c60a2bc0915a"} Mar 21 09:23:57 crc kubenswrapper[4696]: I0321 09:23:57.315735 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t_01e5ce30-73e2-4e41-9d25-6ac3c67f9361/util/0.log" Mar 21 09:23:57 crc kubenswrapper[4696]: I0321 09:23:57.532765 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t_01e5ce30-73e2-4e41-9d25-6ac3c67f9361/pull/0.log" Mar 21 09:23:57 crc kubenswrapper[4696]: I0321 09:23:57.667001 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t_01e5ce30-73e2-4e41-9d25-6ac3c67f9361/util/0.log" Mar 21 09:23:57 crc kubenswrapper[4696]: I0321 09:23:57.673238 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t_01e5ce30-73e2-4e41-9d25-6ac3c67f9361/pull/0.log" Mar 21 09:23:58 crc kubenswrapper[4696]: I0321 09:23:58.249615 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t_01e5ce30-73e2-4e41-9d25-6ac3c67f9361/pull/0.log" Mar 21 09:23:58 crc kubenswrapper[4696]: I0321 09:23:58.263021 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t_01e5ce30-73e2-4e41-9d25-6ac3c67f9361/util/0.log" Mar 21 09:23:58 crc kubenswrapper[4696]: I0321 09:23:58.399492 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t_01e5ce30-73e2-4e41-9d25-6ac3c67f9361/extract/0.log" Mar 21 09:23:58 crc kubenswrapper[4696]: I0321 09:23:58.944042 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6cc65c69fc-m8ffd_530eca0b-da29-4883-ad3a-1faa1563c65d/manager/0.log" Mar 21 09:23:59 crc kubenswrapper[4696]: I0321 09:23:59.435288 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7d559dcdbd-zxgmc_9ce080a1-65e5-4235-b181-448e8987c590/manager/0.log" Mar 21 09:23:59 crc kubenswrapper[4696]: I0321 09:23:59.861231 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-66dd9d474d-qf5kd_58ec1c2d-43a6-4582-9a21-f1083777ffe5/manager/0.log" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.146272 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568084-b55mn"] Mar 21 09:24:00 crc kubenswrapper[4696]: E0321 09:24:00.146683 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a4f80d-c51e-44d3-b823-9f76f1753cbf" containerName="extract-utilities" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.146700 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a4f80d-c51e-44d3-b823-9f76f1753cbf" containerName="extract-utilities" Mar 21 09:24:00 crc kubenswrapper[4696]: E0321 09:24:00.146714 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="681596e5-4b49-43ec-8939-f94736fc1047" containerName="extract-content" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.146721 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="681596e5-4b49-43ec-8939-f94736fc1047" containerName="extract-content" Mar 21 09:24:00 crc kubenswrapper[4696]: E0321 09:24:00.146729 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a4f80d-c51e-44d3-b823-9f76f1753cbf" containerName="registry-server" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.146738 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a4f80d-c51e-44d3-b823-9f76f1753cbf" containerName="registry-server" Mar 21 09:24:00 crc kubenswrapper[4696]: E0321 09:24:00.146750 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="681596e5-4b49-43ec-8939-f94736fc1047" containerName="registry-server" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.146756 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="681596e5-4b49-43ec-8939-f94736fc1047" containerName="registry-server" Mar 21 09:24:00 crc kubenswrapper[4696]: E0321 09:24:00.146769 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60fffaff-5238-4b84-882b-16a0b4c542f6" containerName="oc" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.146774 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="60fffaff-5238-4b84-882b-16a0b4c542f6" containerName="oc" Mar 21 09:24:00 crc kubenswrapper[4696]: E0321 09:24:00.146805 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="681596e5-4b49-43ec-8939-f94736fc1047" containerName="extract-utilities" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.146811 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="681596e5-4b49-43ec-8939-f94736fc1047" containerName="extract-utilities" Mar 21 09:24:00 crc kubenswrapper[4696]: E0321 09:24:00.146915 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a4f80d-c51e-44d3-b823-9f76f1753cbf" containerName="extract-content" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.146921 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a4f80d-c51e-44d3-b823-9f76f1753cbf" containerName="extract-content" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.147101 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="681596e5-4b49-43ec-8939-f94736fc1047" containerName="registry-server" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.147135 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="60fffaff-5238-4b84-882b-16a0b4c542f6" containerName="oc" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.147145 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4a4f80d-c51e-44d3-b823-9f76f1753cbf" containerName="registry-server" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.147896 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568084-b55mn" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.152508 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.152716 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.152894 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.179330 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568084-b55mn"] Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.181460 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-64dc66d669-8685h_9b05c8a9-af90-4497-94f3-8fed890ab3b0/manager/0.log" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.316546 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv9dc\" (UniqueName: \"kubernetes.io/projected/ff83029d-26f8-46b7-b54f-5a8fd3b8598e-kube-api-access-vv9dc\") pod \"auto-csr-approver-29568084-b55mn\" (UID: \"ff83029d-26f8-46b7-b54f-5a8fd3b8598e\") " pod="openshift-infra/auto-csr-approver-29568084-b55mn" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.418051 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv9dc\" (UniqueName: \"kubernetes.io/projected/ff83029d-26f8-46b7-b54f-5a8fd3b8598e-kube-api-access-vv9dc\") pod \"auto-csr-approver-29568084-b55mn\" (UID: \"ff83029d-26f8-46b7-b54f-5a8fd3b8598e\") " pod="openshift-infra/auto-csr-approver-29568084-b55mn" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.447400 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv9dc\" (UniqueName: \"kubernetes.io/projected/ff83029d-26f8-46b7-b54f-5a8fd3b8598e-kube-api-access-vv9dc\") pod \"auto-csr-approver-29568084-b55mn\" (UID: \"ff83029d-26f8-46b7-b54f-5a8fd3b8598e\") " pod="openshift-infra/auto-csr-approver-29568084-b55mn" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.474349 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568084-b55mn" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.525249 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6d77645966-qsd27_3cd62d0b-d740-4bef-ac3c-993b5ee05d40/manager/0.log" Mar 21 09:24:00 crc kubenswrapper[4696]: I0321 09:24:00.997123 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6b77b7676d-vhnjl_74fc6c6a-6eb4-4459-aed9-62073e0472e4/manager/0.log" Mar 21 09:24:01 crc kubenswrapper[4696]: I0321 09:24:01.142924 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5595c7d6ff-7cg24_0048e137-52d0-476b-845a-4c42829de1e2/manager/0.log" Mar 21 09:24:01 crc kubenswrapper[4696]: I0321 09:24:01.284680 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568084-b55mn"] Mar 21 09:24:01 crc kubenswrapper[4696]: I0321 09:24:01.594889 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568084-b55mn" event={"ID":"ff83029d-26f8-46b7-b54f-5a8fd3b8598e","Type":"ContainerStarted","Data":"001e4c0ea56f58d960fa84b0f0d030e48e2e25f5fb7f3ab394e65655b5b5d3f8"} Mar 21 09:24:02 crc kubenswrapper[4696]: I0321 09:24:02.006442 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-76b87776c9-rcz2g_c31077f0-de38-4de7-8421-a5ea7ac3ab84/manager/0.log" Mar 21 09:24:02 crc kubenswrapper[4696]: I0321 09:24:02.243259 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-fbf7bbb96-745cq_8e08ed5d-23b6-4066-a86f-3019f8cb34c8/manager/0.log" Mar 21 09:24:02 crc kubenswrapper[4696]: I0321 09:24:02.342970 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6f5b7bcd4-pzgkp_a4bbdaf9-e480-4cd1-9bf8-304cf5266949/manager/0.log" Mar 21 09:24:02 crc kubenswrapper[4696]: I0321 09:24:02.899970 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6744dd545c-mvqjq_f9ca1637-123c-47eb-8183-5394a0ba6f89/manager/0.log" Mar 21 09:24:03 crc kubenswrapper[4696]: I0321 09:24:03.253568 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-56f74467c6-fzv5q_828a9554-fa0a-4458-9d63-b9e6bd6bde49/manager/0.log" Mar 21 09:24:03 crc kubenswrapper[4696]: I0321 09:24:03.309376 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-bc5c78db9-c9skl_50f56ad0-1409-4619-9b7a-887ba6997874/manager/0.log" Mar 21 09:24:03 crc kubenswrapper[4696]: I0321 09:24:03.638459 4696 generic.go:334] "Generic (PLEG): container finished" podID="ff83029d-26f8-46b7-b54f-5a8fd3b8598e" containerID="fbeec83987fec0cc2ddffebc516606011b51fcf88cc4f7355a71d4e239f53fe2" exitCode=0 Mar 21 09:24:03 crc kubenswrapper[4696]: I0321 09:24:03.638497 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568084-b55mn" event={"ID":"ff83029d-26f8-46b7-b54f-5a8fd3b8598e","Type":"ContainerDied","Data":"fbeec83987fec0cc2ddffebc516606011b51fcf88cc4f7355a71d4e239f53fe2"} Mar 21 09:24:03 crc kubenswrapper[4696]: I0321 09:24:03.711445 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89_026783e3-cff7-40bf-9f26-2d5e8b6006e7/manager/0.log" Mar 21 09:24:03 crc kubenswrapper[4696]: I0321 09:24:03.981355 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-888f66487-rsc92_bfd4f311-01db-41ce-8357-93860d30b442/operator/0.log" Mar 21 09:24:04 crc kubenswrapper[4696]: I0321 09:24:04.403273 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-k24x8_2e3b5f4d-52bd-4d2c-b629-74fc14699179/registry-server/0.log" Mar 21 09:24:04 crc kubenswrapper[4696]: I0321 09:24:04.508569 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-846c4cdcb7-4s5br_1c77ab60-1377-4880-80a9-eb1af2009bca/manager/0.log" Mar 21 09:24:04 crc kubenswrapper[4696]: I0321 09:24:04.807301 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-659fb58c6b-m5bx9_d2672ea4-e1e0-4809-b4e9-04e2cece2179/manager/0.log" Mar 21 09:24:05 crc kubenswrapper[4696]: I0321 09:24:05.438005 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-kvljx_1ff5cfb7-37ad-4f6a-84f1-e8dbdef06084/operator/0.log" Mar 21 09:24:05 crc kubenswrapper[4696]: I0321 09:24:05.625119 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-867f54bc44-n94f2_5fe44a3f-65e2-481b-9d1d-f3b3c24944db/manager/0.log" Mar 21 09:24:05 crc kubenswrapper[4696]: I0321 09:24:05.661202 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568084-b55mn" event={"ID":"ff83029d-26f8-46b7-b54f-5a8fd3b8598e","Type":"ContainerDied","Data":"001e4c0ea56f58d960fa84b0f0d030e48e2e25f5fb7f3ab394e65655b5b5d3f8"} Mar 21 09:24:05 crc kubenswrapper[4696]: I0321 09:24:05.661250 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="001e4c0ea56f58d960fa84b0f0d030e48e2e25f5fb7f3ab394e65655b5b5d3f8" Mar 21 09:24:05 crc kubenswrapper[4696]: I0321 09:24:05.722453 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568084-b55mn" Mar 21 09:24:05 crc kubenswrapper[4696]: I0321 09:24:05.918307 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vv9dc\" (UniqueName: \"kubernetes.io/projected/ff83029d-26f8-46b7-b54f-5a8fd3b8598e-kube-api-access-vv9dc\") pod \"ff83029d-26f8-46b7-b54f-5a8fd3b8598e\" (UID: \"ff83029d-26f8-46b7-b54f-5a8fd3b8598e\") " Mar 21 09:24:05 crc kubenswrapper[4696]: I0321 09:24:05.928014 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff83029d-26f8-46b7-b54f-5a8fd3b8598e-kube-api-access-vv9dc" (OuterVolumeSpecName: "kube-api-access-vv9dc") pod "ff83029d-26f8-46b7-b54f-5a8fd3b8598e" (UID: "ff83029d-26f8-46b7-b54f-5a8fd3b8598e"). InnerVolumeSpecName "kube-api-access-vv9dc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:24:06 crc kubenswrapper[4696]: I0321 09:24:06.022100 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vv9dc\" (UniqueName: \"kubernetes.io/projected/ff83029d-26f8-46b7-b54f-5a8fd3b8598e-kube-api-access-vv9dc\") on node \"crc\" DevicePath \"\"" Mar 21 09:24:06 crc kubenswrapper[4696]: I0321 09:24:06.148209 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-8467ccb4c8-zk58x_46974657-c404-4081-a4ef-5fc99237a02f/manager/0.log" Mar 21 09:24:06 crc kubenswrapper[4696]: I0321 09:24:06.392622 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-b7fdfc6b-hrhxc_fbbc65c6-7696-4795-8b4b-7d9c7e315bf0/manager/0.log" Mar 21 09:24:06 crc kubenswrapper[4696]: I0321 09:24:06.557685 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-74d6f7b5c-5x9b6_f593acbc-0a73-41ed-a949-02a3556dfc70/manager/0.log" Mar 21 09:24:06 crc kubenswrapper[4696]: I0321 09:24:06.569584 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-69d78c9949-5whvx_0ecf9ee3-d5c9-4677-b264-0ddae28df715/manager/0.log" Mar 21 09:24:06 crc kubenswrapper[4696]: I0321 09:24:06.672116 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568084-b55mn" Mar 21 09:24:06 crc kubenswrapper[4696]: I0321 09:24:06.748042 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5cfd84c587-ldj9m_53a9e616-007d-4c0c-8b0a-799d7188d9ab/manager/0.log" Mar 21 09:24:06 crc kubenswrapper[4696]: I0321 09:24:06.808377 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568078-gprd7"] Mar 21 09:24:06 crc kubenswrapper[4696]: I0321 09:24:06.835016 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568078-gprd7"] Mar 21 09:24:08 crc kubenswrapper[4696]: I0321 09:24:08.544882 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07bdef47-c372-4b8f-a093-3a31c5231fb9" path="/var/lib/kubelet/pods/07bdef47-c372-4b8f-a093-3a31c5231fb9/volumes" Mar 21 09:24:41 crc kubenswrapper[4696]: I0321 09:24:41.757357 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-5mlz9_6769f68b-1f58-48e4-8076-14fcdb9c8036/control-plane-machine-set-operator/0.log" Mar 21 09:24:42 crc kubenswrapper[4696]: I0321 09:24:42.040172 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-h5tfm_f62330e9-cdc5-469b-82ba-3364ce173686/kube-rbac-proxy/0.log" Mar 21 09:24:42 crc kubenswrapper[4696]: I0321 09:24:42.170171 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-h5tfm_f62330e9-cdc5-469b-82ba-3364ce173686/machine-api-operator/0.log" Mar 21 09:24:53 crc kubenswrapper[4696]: I0321 09:24:53.132712 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7tjcn"] Mar 21 09:24:53 crc kubenswrapper[4696]: E0321 09:24:53.133883 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff83029d-26f8-46b7-b54f-5a8fd3b8598e" containerName="oc" Mar 21 09:24:53 crc kubenswrapper[4696]: I0321 09:24:53.133897 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff83029d-26f8-46b7-b54f-5a8fd3b8598e" containerName="oc" Mar 21 09:24:53 crc kubenswrapper[4696]: I0321 09:24:53.134106 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff83029d-26f8-46b7-b54f-5a8fd3b8598e" containerName="oc" Mar 21 09:24:53 crc kubenswrapper[4696]: I0321 09:24:53.136011 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tjcn" Mar 21 09:24:53 crc kubenswrapper[4696]: I0321 09:24:53.162024 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7tjcn"] Mar 21 09:24:53 crc kubenswrapper[4696]: I0321 09:24:53.191975 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcv5r\" (UniqueName: \"kubernetes.io/projected/fceab736-45b8-4d83-bee0-36e3aaf711c1-kube-api-access-zcv5r\") pod \"redhat-operators-7tjcn\" (UID: \"fceab736-45b8-4d83-bee0-36e3aaf711c1\") " pod="openshift-marketplace/redhat-operators-7tjcn" Mar 21 09:24:53 crc kubenswrapper[4696]: I0321 09:24:53.192053 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fceab736-45b8-4d83-bee0-36e3aaf711c1-catalog-content\") pod \"redhat-operators-7tjcn\" (UID: \"fceab736-45b8-4d83-bee0-36e3aaf711c1\") " pod="openshift-marketplace/redhat-operators-7tjcn" Mar 21 09:24:53 crc kubenswrapper[4696]: I0321 09:24:53.192081 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fceab736-45b8-4d83-bee0-36e3aaf711c1-utilities\") pod \"redhat-operators-7tjcn\" (UID: \"fceab736-45b8-4d83-bee0-36e3aaf711c1\") " pod="openshift-marketplace/redhat-operators-7tjcn" Mar 21 09:24:53 crc kubenswrapper[4696]: I0321 09:24:53.296123 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcv5r\" (UniqueName: \"kubernetes.io/projected/fceab736-45b8-4d83-bee0-36e3aaf711c1-kube-api-access-zcv5r\") pod \"redhat-operators-7tjcn\" (UID: \"fceab736-45b8-4d83-bee0-36e3aaf711c1\") " pod="openshift-marketplace/redhat-operators-7tjcn" Mar 21 09:24:53 crc kubenswrapper[4696]: I0321 09:24:53.296205 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fceab736-45b8-4d83-bee0-36e3aaf711c1-catalog-content\") pod \"redhat-operators-7tjcn\" (UID: \"fceab736-45b8-4d83-bee0-36e3aaf711c1\") " pod="openshift-marketplace/redhat-operators-7tjcn" Mar 21 09:24:53 crc kubenswrapper[4696]: I0321 09:24:53.296237 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fceab736-45b8-4d83-bee0-36e3aaf711c1-utilities\") pod \"redhat-operators-7tjcn\" (UID: \"fceab736-45b8-4d83-bee0-36e3aaf711c1\") " pod="openshift-marketplace/redhat-operators-7tjcn" Mar 21 09:24:53 crc kubenswrapper[4696]: I0321 09:24:53.296655 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fceab736-45b8-4d83-bee0-36e3aaf711c1-catalog-content\") pod \"redhat-operators-7tjcn\" (UID: \"fceab736-45b8-4d83-bee0-36e3aaf711c1\") " pod="openshift-marketplace/redhat-operators-7tjcn" Mar 21 09:24:53 crc kubenswrapper[4696]: I0321 09:24:53.296686 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fceab736-45b8-4d83-bee0-36e3aaf711c1-utilities\") pod \"redhat-operators-7tjcn\" (UID: \"fceab736-45b8-4d83-bee0-36e3aaf711c1\") " pod="openshift-marketplace/redhat-operators-7tjcn" Mar 21 09:24:53 crc kubenswrapper[4696]: I0321 09:24:53.316624 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcv5r\" (UniqueName: \"kubernetes.io/projected/fceab736-45b8-4d83-bee0-36e3aaf711c1-kube-api-access-zcv5r\") pod \"redhat-operators-7tjcn\" (UID: \"fceab736-45b8-4d83-bee0-36e3aaf711c1\") " pod="openshift-marketplace/redhat-operators-7tjcn" Mar 21 09:24:53 crc kubenswrapper[4696]: I0321 09:24:53.467376 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tjcn" Mar 21 09:24:54 crc kubenswrapper[4696]: I0321 09:24:54.240666 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7tjcn"] Mar 21 09:24:55 crc kubenswrapper[4696]: I0321 09:24:55.134670 4696 generic.go:334] "Generic (PLEG): container finished" podID="fceab736-45b8-4d83-bee0-36e3aaf711c1" containerID="a5d3282ad5299990d6e28448bfb90921fd7970539b3064bea640c029b1bd6c0c" exitCode=0 Mar 21 09:24:55 crc kubenswrapper[4696]: I0321 09:24:55.134768 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tjcn" event={"ID":"fceab736-45b8-4d83-bee0-36e3aaf711c1","Type":"ContainerDied","Data":"a5d3282ad5299990d6e28448bfb90921fd7970539b3064bea640c029b1bd6c0c"} Mar 21 09:24:55 crc kubenswrapper[4696]: I0321 09:24:55.135232 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tjcn" event={"ID":"fceab736-45b8-4d83-bee0-36e3aaf711c1","Type":"ContainerStarted","Data":"1017ec7efb2402ac290f200fd6093ad23452100b6a5e8f24e77eb8c96616c3bd"} Mar 21 09:24:55 crc kubenswrapper[4696]: I0321 09:24:55.136886 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 09:24:56 crc kubenswrapper[4696]: I0321 09:24:56.145006 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tjcn" event={"ID":"fceab736-45b8-4d83-bee0-36e3aaf711c1","Type":"ContainerStarted","Data":"93859136fb7da5f3442cc3109d63a7871e0cc7ce3133e72e0af8796256ef1669"} Mar 21 09:25:02 crc kubenswrapper[4696]: I0321 09:25:02.195962 4696 generic.go:334] "Generic (PLEG): container finished" podID="fceab736-45b8-4d83-bee0-36e3aaf711c1" containerID="93859136fb7da5f3442cc3109d63a7871e0cc7ce3133e72e0af8796256ef1669" exitCode=0 Mar 21 09:25:02 crc kubenswrapper[4696]: I0321 09:25:02.196060 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tjcn" event={"ID":"fceab736-45b8-4d83-bee0-36e3aaf711c1","Type":"ContainerDied","Data":"93859136fb7da5f3442cc3109d63a7871e0cc7ce3133e72e0af8796256ef1669"} Mar 21 09:25:02 crc kubenswrapper[4696]: I0321 09:25:02.991626 4696 scope.go:117] "RemoveContainer" containerID="11da25fec5cad12f50d5cd8a74c193bf96237eb669b12b759bc6a941e076c8f5" Mar 21 09:25:03 crc kubenswrapper[4696]: I0321 09:25:03.209966 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tjcn" event={"ID":"fceab736-45b8-4d83-bee0-36e3aaf711c1","Type":"ContainerStarted","Data":"1e151fc4ede4eac777629dc814aa17ce3c7dafb5351d72f7a6a1ef307965f2bc"} Mar 21 09:25:03 crc kubenswrapper[4696]: I0321 09:25:03.237131 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7tjcn" podStartSLOduration=2.6634199880000002 podStartE2EDuration="10.237106643s" podCreationTimestamp="2026-03-21 09:24:53 +0000 UTC" firstStartedPulling="2026-03-21 09:24:55.136654699 +0000 UTC m=+3429.257535412" lastFinishedPulling="2026-03-21 09:25:02.710341354 +0000 UTC m=+3436.831222067" observedRunningTime="2026-03-21 09:25:03.232524511 +0000 UTC m=+3437.353405224" watchObservedRunningTime="2026-03-21 09:25:03.237106643 +0000 UTC m=+3437.357987366" Mar 21 09:25:03 crc kubenswrapper[4696]: I0321 09:25:03.467850 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7tjcn" Mar 21 09:25:03 crc kubenswrapper[4696]: I0321 09:25:03.467931 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7tjcn" Mar 21 09:25:04 crc kubenswrapper[4696]: I0321 09:25:04.524615 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7tjcn" podUID="fceab736-45b8-4d83-bee0-36e3aaf711c1" containerName="registry-server" probeResult="failure" output=< Mar 21 09:25:04 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Mar 21 09:25:04 crc kubenswrapper[4696]: > Mar 21 09:25:05 crc kubenswrapper[4696]: I0321 09:25:05.374469 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-xbjnb_f6174e0b-3b88-4452-ab2f-89905f5d09a8/cert-manager-controller/0.log" Mar 21 09:25:05 crc kubenswrapper[4696]: I0321 09:25:05.943252 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-f85xz_5a3acbc7-10ae-4287-9d73-15b97d7f0d71/cert-manager-cainjector/0.log" Mar 21 09:25:05 crc kubenswrapper[4696]: I0321 09:25:05.997987 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-rp8tc_e0d6313d-e133-464b-b5ea-33699c8280c0/cert-manager-webhook/0.log" Mar 21 09:25:13 crc kubenswrapper[4696]: I0321 09:25:13.544339 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7tjcn" Mar 21 09:25:13 crc kubenswrapper[4696]: I0321 09:25:13.602919 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7tjcn" Mar 21 09:25:13 crc kubenswrapper[4696]: I0321 09:25:13.781594 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7tjcn"] Mar 21 09:25:15 crc kubenswrapper[4696]: I0321 09:25:15.321175 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7tjcn" podUID="fceab736-45b8-4d83-bee0-36e3aaf711c1" containerName="registry-server" containerID="cri-o://1e151fc4ede4eac777629dc814aa17ce3c7dafb5351d72f7a6a1ef307965f2bc" gracePeriod=2 Mar 21 09:25:16 crc kubenswrapper[4696]: I0321 09:25:16.334783 4696 generic.go:334] "Generic (PLEG): container finished" podID="fceab736-45b8-4d83-bee0-36e3aaf711c1" containerID="1e151fc4ede4eac777629dc814aa17ce3c7dafb5351d72f7a6a1ef307965f2bc" exitCode=0 Mar 21 09:25:16 crc kubenswrapper[4696]: I0321 09:25:16.334913 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tjcn" event={"ID":"fceab736-45b8-4d83-bee0-36e3aaf711c1","Type":"ContainerDied","Data":"1e151fc4ede4eac777629dc814aa17ce3c7dafb5351d72f7a6a1ef307965f2bc"} Mar 21 09:25:16 crc kubenswrapper[4696]: I0321 09:25:16.497283 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tjcn" Mar 21 09:25:16 crc kubenswrapper[4696]: I0321 09:25:16.595271 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fceab736-45b8-4d83-bee0-36e3aaf711c1-utilities\") pod \"fceab736-45b8-4d83-bee0-36e3aaf711c1\" (UID: \"fceab736-45b8-4d83-bee0-36e3aaf711c1\") " Mar 21 09:25:16 crc kubenswrapper[4696]: I0321 09:25:16.595483 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fceab736-45b8-4d83-bee0-36e3aaf711c1-catalog-content\") pod \"fceab736-45b8-4d83-bee0-36e3aaf711c1\" (UID: \"fceab736-45b8-4d83-bee0-36e3aaf711c1\") " Mar 21 09:25:16 crc kubenswrapper[4696]: I0321 09:25:16.595611 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcv5r\" (UniqueName: \"kubernetes.io/projected/fceab736-45b8-4d83-bee0-36e3aaf711c1-kube-api-access-zcv5r\") pod \"fceab736-45b8-4d83-bee0-36e3aaf711c1\" (UID: \"fceab736-45b8-4d83-bee0-36e3aaf711c1\") " Mar 21 09:25:16 crc kubenswrapper[4696]: I0321 09:25:16.596700 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fceab736-45b8-4d83-bee0-36e3aaf711c1-utilities" (OuterVolumeSpecName: "utilities") pod "fceab736-45b8-4d83-bee0-36e3aaf711c1" (UID: "fceab736-45b8-4d83-bee0-36e3aaf711c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:25:16 crc kubenswrapper[4696]: I0321 09:25:16.614451 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fceab736-45b8-4d83-bee0-36e3aaf711c1-kube-api-access-zcv5r" (OuterVolumeSpecName: "kube-api-access-zcv5r") pod "fceab736-45b8-4d83-bee0-36e3aaf711c1" (UID: "fceab736-45b8-4d83-bee0-36e3aaf711c1"). InnerVolumeSpecName "kube-api-access-zcv5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:25:16 crc kubenswrapper[4696]: I0321 09:25:16.704161 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fceab736-45b8-4d83-bee0-36e3aaf711c1-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 09:25:16 crc kubenswrapper[4696]: I0321 09:25:16.704200 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcv5r\" (UniqueName: \"kubernetes.io/projected/fceab736-45b8-4d83-bee0-36e3aaf711c1-kube-api-access-zcv5r\") on node \"crc\" DevicePath \"\"" Mar 21 09:25:16 crc kubenswrapper[4696]: I0321 09:25:16.784219 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fceab736-45b8-4d83-bee0-36e3aaf711c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fceab736-45b8-4d83-bee0-36e3aaf711c1" (UID: "fceab736-45b8-4d83-bee0-36e3aaf711c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:25:16 crc kubenswrapper[4696]: I0321 09:25:16.806594 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fceab736-45b8-4d83-bee0-36e3aaf711c1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 09:25:17 crc kubenswrapper[4696]: I0321 09:25:17.345559 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tjcn" event={"ID":"fceab736-45b8-4d83-bee0-36e3aaf711c1","Type":"ContainerDied","Data":"1017ec7efb2402ac290f200fd6093ad23452100b6a5e8f24e77eb8c96616c3bd"} Mar 21 09:25:17 crc kubenswrapper[4696]: I0321 09:25:17.345624 4696 scope.go:117] "RemoveContainer" containerID="1e151fc4ede4eac777629dc814aa17ce3c7dafb5351d72f7a6a1ef307965f2bc" Mar 21 09:25:17 crc kubenswrapper[4696]: I0321 09:25:17.346451 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tjcn" Mar 21 09:25:17 crc kubenswrapper[4696]: I0321 09:25:17.366856 4696 scope.go:117] "RemoveContainer" containerID="93859136fb7da5f3442cc3109d63a7871e0cc7ce3133e72e0af8796256ef1669" Mar 21 09:25:17 crc kubenswrapper[4696]: I0321 09:25:17.400946 4696 scope.go:117] "RemoveContainer" containerID="a5d3282ad5299990d6e28448bfb90921fd7970539b3064bea640c029b1bd6c0c" Mar 21 09:25:17 crc kubenswrapper[4696]: I0321 09:25:17.420683 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7tjcn"] Mar 21 09:25:17 crc kubenswrapper[4696]: I0321 09:25:17.462513 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7tjcn"] Mar 21 09:25:18 crc kubenswrapper[4696]: I0321 09:25:18.550448 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fceab736-45b8-4d83-bee0-36e3aaf711c1" path="/var/lib/kubelet/pods/fceab736-45b8-4d83-bee0-36e3aaf711c1/volumes" Mar 21 09:25:29 crc kubenswrapper[4696]: I0321 09:25:29.380290 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-k9rdf_81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160/nmstate-console-plugin/0.log" Mar 21 09:25:29 crc kubenswrapper[4696]: I0321 09:25:29.701279 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-r8wnr_be6dcd14-1d77-4c67-b527-29e7f46efc3e/nmstate-handler/0.log" Mar 21 09:25:29 crc kubenswrapper[4696]: I0321 09:25:29.845265 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-q8tc5_078ef7e2-dd73-486e-9213-804f5dbd26b2/kube-rbac-proxy/0.log" Mar 21 09:25:29 crc kubenswrapper[4696]: I0321 09:25:29.915907 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-q8tc5_078ef7e2-dd73-486e-9213-804f5dbd26b2/nmstate-metrics/0.log" Mar 21 09:25:30 crc kubenswrapper[4696]: I0321 09:25:30.167336 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-zc76d_3e70f161-07ef-4ed3-90be-6930cfb666a4/nmstate-operator/0.log" Mar 21 09:25:30 crc kubenswrapper[4696]: I0321 09:25:30.217146 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-q8s95_a351e45b-b9be-4898-8e07-c63cdd497245/nmstate-webhook/0.log" Mar 21 09:25:30 crc kubenswrapper[4696]: I0321 09:25:30.340742 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:25:30 crc kubenswrapper[4696]: I0321 09:25:30.340793 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:25:50 crc kubenswrapper[4696]: I0321 09:25:50.881597 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5685547f78-5x2wx_75d1fdab-70ac-46df-8b60-3e2f75985686/kube-rbac-proxy/0.log" Mar 21 09:25:50 crc kubenswrapper[4696]: I0321 09:25:50.923937 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5685547f78-5x2wx_75d1fdab-70ac-46df-8b60-3e2f75985686/manager/0.log" Mar 21 09:26:00 crc kubenswrapper[4696]: I0321 09:26:00.143554 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568086-dx7lf"] Mar 21 09:26:00 crc kubenswrapper[4696]: E0321 09:26:00.144590 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fceab736-45b8-4d83-bee0-36e3aaf711c1" containerName="registry-server" Mar 21 09:26:00 crc kubenswrapper[4696]: I0321 09:26:00.144603 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="fceab736-45b8-4d83-bee0-36e3aaf711c1" containerName="registry-server" Mar 21 09:26:00 crc kubenswrapper[4696]: E0321 09:26:00.144628 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fceab736-45b8-4d83-bee0-36e3aaf711c1" containerName="extract-content" Mar 21 09:26:00 crc kubenswrapper[4696]: I0321 09:26:00.144634 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="fceab736-45b8-4d83-bee0-36e3aaf711c1" containerName="extract-content" Mar 21 09:26:00 crc kubenswrapper[4696]: E0321 09:26:00.144645 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fceab736-45b8-4d83-bee0-36e3aaf711c1" containerName="extract-utilities" Mar 21 09:26:00 crc kubenswrapper[4696]: I0321 09:26:00.144651 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="fceab736-45b8-4d83-bee0-36e3aaf711c1" containerName="extract-utilities" Mar 21 09:26:00 crc kubenswrapper[4696]: I0321 09:26:00.144873 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="fceab736-45b8-4d83-bee0-36e3aaf711c1" containerName="registry-server" Mar 21 09:26:00 crc kubenswrapper[4696]: I0321 09:26:00.145582 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568086-dx7lf" Mar 21 09:26:00 crc kubenswrapper[4696]: I0321 09:26:00.148057 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:26:00 crc kubenswrapper[4696]: I0321 09:26:00.148069 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:26:00 crc kubenswrapper[4696]: I0321 09:26:00.148669 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:26:00 crc kubenswrapper[4696]: I0321 09:26:00.156960 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568086-dx7lf"] Mar 21 09:26:00 crc kubenswrapper[4696]: I0321 09:26:00.316916 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t6tf\" (UniqueName: \"kubernetes.io/projected/57992ac0-359d-43ba-b9f3-d5e50e96cebd-kube-api-access-5t6tf\") pod \"auto-csr-approver-29568086-dx7lf\" (UID: \"57992ac0-359d-43ba-b9f3-d5e50e96cebd\") " pod="openshift-infra/auto-csr-approver-29568086-dx7lf" Mar 21 09:26:00 crc kubenswrapper[4696]: I0321 09:26:00.340936 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:26:00 crc kubenswrapper[4696]: I0321 09:26:00.341006 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:26:00 crc kubenswrapper[4696]: I0321 09:26:00.419284 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t6tf\" (UniqueName: \"kubernetes.io/projected/57992ac0-359d-43ba-b9f3-d5e50e96cebd-kube-api-access-5t6tf\") pod \"auto-csr-approver-29568086-dx7lf\" (UID: \"57992ac0-359d-43ba-b9f3-d5e50e96cebd\") " pod="openshift-infra/auto-csr-approver-29568086-dx7lf" Mar 21 09:26:00 crc kubenswrapper[4696]: I0321 09:26:00.447469 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t6tf\" (UniqueName: \"kubernetes.io/projected/57992ac0-359d-43ba-b9f3-d5e50e96cebd-kube-api-access-5t6tf\") pod \"auto-csr-approver-29568086-dx7lf\" (UID: \"57992ac0-359d-43ba-b9f3-d5e50e96cebd\") " pod="openshift-infra/auto-csr-approver-29568086-dx7lf" Mar 21 09:26:00 crc kubenswrapper[4696]: I0321 09:26:00.473712 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568086-dx7lf" Mar 21 09:26:01 crc kubenswrapper[4696]: I0321 09:26:01.229492 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568086-dx7lf"] Mar 21 09:26:01 crc kubenswrapper[4696]: I0321 09:26:01.763331 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568086-dx7lf" event={"ID":"57992ac0-359d-43ba-b9f3-d5e50e96cebd","Type":"ContainerStarted","Data":"976a58052f2169b262578e4e3c5b0cf719066861b601979e0018e2520eac5ff3"} Mar 21 09:26:02 crc kubenswrapper[4696]: I0321 09:26:02.774471 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568086-dx7lf" event={"ID":"57992ac0-359d-43ba-b9f3-d5e50e96cebd","Type":"ContainerStarted","Data":"c3af5de76bf59ba40cbb8323b2356ac81f1ac31ab6813d2cf714784f52f68387"} Mar 21 09:26:02 crc kubenswrapper[4696]: I0321 09:26:02.793030 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29568086-dx7lf" podStartSLOduration=1.758702801 podStartE2EDuration="2.792988601s" podCreationTimestamp="2026-03-21 09:26:00 +0000 UTC" firstStartedPulling="2026-03-21 09:26:01.226068796 +0000 UTC m=+3495.346949509" lastFinishedPulling="2026-03-21 09:26:02.260354596 +0000 UTC m=+3496.381235309" observedRunningTime="2026-03-21 09:26:02.786922677 +0000 UTC m=+3496.907803390" watchObservedRunningTime="2026-03-21 09:26:02.792988601 +0000 UTC m=+3496.913869304" Mar 21 09:26:03 crc kubenswrapper[4696]: I0321 09:26:03.786389 4696 generic.go:334] "Generic (PLEG): container finished" podID="57992ac0-359d-43ba-b9f3-d5e50e96cebd" containerID="c3af5de76bf59ba40cbb8323b2356ac81f1ac31ab6813d2cf714784f52f68387" exitCode=0 Mar 21 09:26:03 crc kubenswrapper[4696]: I0321 09:26:03.786533 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568086-dx7lf" event={"ID":"57992ac0-359d-43ba-b9f3-d5e50e96cebd","Type":"ContainerDied","Data":"c3af5de76bf59ba40cbb8323b2356ac81f1ac31ab6813d2cf714784f52f68387"} Mar 21 09:26:05 crc kubenswrapper[4696]: I0321 09:26:05.805162 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568086-dx7lf" event={"ID":"57992ac0-359d-43ba-b9f3-d5e50e96cebd","Type":"ContainerDied","Data":"976a58052f2169b262578e4e3c5b0cf719066861b601979e0018e2520eac5ff3"} Mar 21 09:26:05 crc kubenswrapper[4696]: I0321 09:26:05.805605 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="976a58052f2169b262578e4e3c5b0cf719066861b601979e0018e2520eac5ff3" Mar 21 09:26:05 crc kubenswrapper[4696]: I0321 09:26:05.893017 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568086-dx7lf" Mar 21 09:26:06 crc kubenswrapper[4696]: I0321 09:26:06.043727 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5t6tf\" (UniqueName: \"kubernetes.io/projected/57992ac0-359d-43ba-b9f3-d5e50e96cebd-kube-api-access-5t6tf\") pod \"57992ac0-359d-43ba-b9f3-d5e50e96cebd\" (UID: \"57992ac0-359d-43ba-b9f3-d5e50e96cebd\") " Mar 21 09:26:06 crc kubenswrapper[4696]: I0321 09:26:06.052569 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57992ac0-359d-43ba-b9f3-d5e50e96cebd-kube-api-access-5t6tf" (OuterVolumeSpecName: "kube-api-access-5t6tf") pod "57992ac0-359d-43ba-b9f3-d5e50e96cebd" (UID: "57992ac0-359d-43ba-b9f3-d5e50e96cebd"). InnerVolumeSpecName "kube-api-access-5t6tf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:26:06 crc kubenswrapper[4696]: I0321 09:26:06.147087 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5t6tf\" (UniqueName: \"kubernetes.io/projected/57992ac0-359d-43ba-b9f3-d5e50e96cebd-kube-api-access-5t6tf\") on node \"crc\" DevicePath \"\"" Mar 21 09:26:06 crc kubenswrapper[4696]: I0321 09:26:06.822492 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568086-dx7lf" Mar 21 09:26:06 crc kubenswrapper[4696]: I0321 09:26:06.968586 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568080-82ltv"] Mar 21 09:26:06 crc kubenswrapper[4696]: I0321 09:26:06.986039 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568080-82ltv"] Mar 21 09:26:08 crc kubenswrapper[4696]: I0321 09:26:08.546162 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="640729b7-7edd-4af6-b156-811c9aaf7574" path="/var/lib/kubelet/pods/640729b7-7edd-4af6-b156-811c9aaf7574/volumes" Mar 21 09:26:11 crc kubenswrapper[4696]: I0321 09:26:11.709917 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-8ff7d675-rs27t_527ae575-3dfd-4f0a-9af9-63668af676d3/prometheus-operator/0.log" Mar 21 09:26:12 crc kubenswrapper[4696]: I0321 09:26:12.008444 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-db48b98cb-78qsn_474e7a66-fb1d-448c-8297-b511ec7b6e7c/prometheus-operator-admission-webhook/0.log" Mar 21 09:26:12 crc kubenswrapper[4696]: I0321 09:26:12.147998 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-db48b98cb-l8q89_d99a53b9-d503-4410-9c7c-807fca5276bc/prometheus-operator-admission-webhook/0.log" Mar 21 09:26:12 crc kubenswrapper[4696]: I0321 09:26:12.342978 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-6dd7dd855f-sp4gg_e864a9cc-e270-4238-841b-77d0abad085d/operator/0.log" Mar 21 09:26:12 crc kubenswrapper[4696]: I0321 09:26:12.414926 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-64bcb4cf6f-jjk5b_127ee197-aa20-41d3-ac26-e8aead566b64/perses-operator/0.log" Mar 21 09:26:30 crc kubenswrapper[4696]: I0321 09:26:30.340914 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:26:30 crc kubenswrapper[4696]: I0321 09:26:30.341491 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:26:30 crc kubenswrapper[4696]: I0321 09:26:30.341539 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 09:26:30 crc kubenswrapper[4696]: I0321 09:26:30.342365 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aa5a03c21bdd6334cc43af15b5fd004e41076b2d7b35f63b8f51c60a2bc0915a"} pod="openshift-machine-config-operator/machine-config-daemon-z7srw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 09:26:30 crc kubenswrapper[4696]: I0321 09:26:30.342419 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" containerID="cri-o://aa5a03c21bdd6334cc43af15b5fd004e41076b2d7b35f63b8f51c60a2bc0915a" gracePeriod=600 Mar 21 09:26:30 crc kubenswrapper[4696]: E0321 09:26:30.607790 4696 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddaaf227a_2305_495e_8495_a280abcd8e10.slice/crio-conmon-aa5a03c21bdd6334cc43af15b5fd004e41076b2d7b35f63b8f51c60a2bc0915a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddaaf227a_2305_495e_8495_a280abcd8e10.slice/crio-aa5a03c21bdd6334cc43af15b5fd004e41076b2d7b35f63b8f51c60a2bc0915a.scope\": RecentStats: unable to find data in memory cache]" Mar 21 09:26:31 crc kubenswrapper[4696]: I0321 09:26:31.036138 4696 generic.go:334] "Generic (PLEG): container finished" podID="daaf227a-2305-495e-8495-a280abcd8e10" containerID="aa5a03c21bdd6334cc43af15b5fd004e41076b2d7b35f63b8f51c60a2bc0915a" exitCode=0 Mar 21 09:26:31 crc kubenswrapper[4696]: I0321 09:26:31.036179 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerDied","Data":"aa5a03c21bdd6334cc43af15b5fd004e41076b2d7b35f63b8f51c60a2bc0915a"} Mar 21 09:26:31 crc kubenswrapper[4696]: I0321 09:26:31.036204 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerStarted","Data":"32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b"} Mar 21 09:26:31 crc kubenswrapper[4696]: I0321 09:26:31.036218 4696 scope.go:117] "RemoveContainer" containerID="5128f38964599f9b55adaec586e600f88289b4a555fe534f73d1d8a8ad3a906c" Mar 21 09:26:35 crc kubenswrapper[4696]: I0321 09:26:35.084384 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-lzh5g_3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c/kube-rbac-proxy/0.log" Mar 21 09:26:35 crc kubenswrapper[4696]: I0321 09:26:35.391338 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-lzh5g_3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c/controller/0.log" Mar 21 09:26:35 crc kubenswrapper[4696]: I0321 09:26:35.569730 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-frr-files/0.log" Mar 21 09:26:35 crc kubenswrapper[4696]: I0321 09:26:35.834749 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-frr-files/0.log" Mar 21 09:26:35 crc kubenswrapper[4696]: I0321 09:26:35.871532 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-metrics/0.log" Mar 21 09:26:35 crc kubenswrapper[4696]: I0321 09:26:35.937703 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-reloader/0.log" Mar 21 09:26:35 crc kubenswrapper[4696]: I0321 09:26:35.981379 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-reloader/0.log" Mar 21 09:26:36 crc kubenswrapper[4696]: I0321 09:26:36.202399 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-frr-files/0.log" Mar 21 09:26:36 crc kubenswrapper[4696]: I0321 09:26:36.242509 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-metrics/0.log" Mar 21 09:26:36 crc kubenswrapper[4696]: I0321 09:26:36.282556 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-reloader/0.log" Mar 21 09:26:36 crc kubenswrapper[4696]: I0321 09:26:36.289649 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-metrics/0.log" Mar 21 09:26:36 crc kubenswrapper[4696]: I0321 09:26:36.497863 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-metrics/0.log" Mar 21 09:26:36 crc kubenswrapper[4696]: I0321 09:26:36.515407 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-frr-files/0.log" Mar 21 09:26:36 crc kubenswrapper[4696]: I0321 09:26:36.562116 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-reloader/0.log" Mar 21 09:26:36 crc kubenswrapper[4696]: I0321 09:26:36.649611 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/controller/0.log" Mar 21 09:26:36 crc kubenswrapper[4696]: I0321 09:26:36.924287 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/kube-rbac-proxy/0.log" Mar 21 09:26:36 crc kubenswrapper[4696]: I0321 09:26:36.926105 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/frr-metrics/0.log" Mar 21 09:26:36 crc kubenswrapper[4696]: I0321 09:26:36.949762 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/kube-rbac-proxy-frr/0.log" Mar 21 09:26:37 crc kubenswrapper[4696]: I0321 09:26:37.346172 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/reloader/0.log" Mar 21 09:26:37 crc kubenswrapper[4696]: I0321 09:26:37.395767 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-gk8sz_69b8f173-b1f8-45e3-a6ae-ffb987fc709e/frr-k8s-webhook-server/0.log" Mar 21 09:26:37 crc kubenswrapper[4696]: I0321 09:26:37.667399 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-78876797bd-qsnl8_5fe21876-0bc8-489f-b862-6d48be898f5b/manager/0.log" Mar 21 09:26:37 crc kubenswrapper[4696]: I0321 09:26:37.883413 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7749f78475-cjlsc_dd16e737-b1c3-451b-a87c-90ec18d32dd3/webhook-server/0.log" Mar 21 09:26:38 crc kubenswrapper[4696]: I0321 09:26:38.361370 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zfggl_d8c2e94f-7eff-41dd-8e91-d6301ac9b84a/kube-rbac-proxy/0.log" Mar 21 09:26:38 crc kubenswrapper[4696]: I0321 09:26:38.468477 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/frr/0.log" Mar 21 09:26:38 crc kubenswrapper[4696]: I0321 09:26:38.636186 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zfggl_d8c2e94f-7eff-41dd-8e91-d6301ac9b84a/speaker/0.log" Mar 21 09:26:59 crc kubenswrapper[4696]: I0321 09:26:59.594386 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8_f46ecbe7-f1ca-4d24-975f-487b15fb2707/util/0.log" Mar 21 09:26:59 crc kubenswrapper[4696]: I0321 09:26:59.846651 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8_f46ecbe7-f1ca-4d24-975f-487b15fb2707/util/0.log" Mar 21 09:26:59 crc kubenswrapper[4696]: I0321 09:26:59.903572 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8_f46ecbe7-f1ca-4d24-975f-487b15fb2707/pull/0.log" Mar 21 09:26:59 crc kubenswrapper[4696]: I0321 09:26:59.957033 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8_f46ecbe7-f1ca-4d24-975f-487b15fb2707/pull/0.log" Mar 21 09:27:00 crc kubenswrapper[4696]: I0321 09:27:00.301868 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8_f46ecbe7-f1ca-4d24-975f-487b15fb2707/pull/0.log" Mar 21 09:27:00 crc kubenswrapper[4696]: I0321 09:27:00.326646 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8_f46ecbe7-f1ca-4d24-975f-487b15fb2707/extract/0.log" Mar 21 09:27:00 crc kubenswrapper[4696]: I0321 09:27:00.403099 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8_f46ecbe7-f1ca-4d24-975f-487b15fb2707/util/0.log" Mar 21 09:27:00 crc kubenswrapper[4696]: I0321 09:27:00.606363 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9_6ed1f144-c0f4-470e-9bc2-fc4ea36bc158/util/0.log" Mar 21 09:27:00 crc kubenswrapper[4696]: I0321 09:27:00.854440 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9_6ed1f144-c0f4-470e-9bc2-fc4ea36bc158/util/0.log" Mar 21 09:27:00 crc kubenswrapper[4696]: I0321 09:27:00.942172 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9_6ed1f144-c0f4-470e-9bc2-fc4ea36bc158/pull/0.log" Mar 21 09:27:00 crc kubenswrapper[4696]: I0321 09:27:00.975760 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9_6ed1f144-c0f4-470e-9bc2-fc4ea36bc158/pull/0.log" Mar 21 09:27:01 crc kubenswrapper[4696]: I0321 09:27:01.225280 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9_6ed1f144-c0f4-470e-9bc2-fc4ea36bc158/util/0.log" Mar 21 09:27:01 crc kubenswrapper[4696]: I0321 09:27:01.293617 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9_6ed1f144-c0f4-470e-9bc2-fc4ea36bc158/pull/0.log" Mar 21 09:27:01 crc kubenswrapper[4696]: I0321 09:27:01.301287 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9_6ed1f144-c0f4-470e-9bc2-fc4ea36bc158/extract/0.log" Mar 21 09:27:01 crc kubenswrapper[4696]: I0321 09:27:01.571897 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5_e4b15b92-b219-4ad6-a9ab-7d05a7130c92/util/0.log" Mar 21 09:27:01 crc kubenswrapper[4696]: I0321 09:27:01.908863 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5_e4b15b92-b219-4ad6-a9ab-7d05a7130c92/util/0.log" Mar 21 09:27:01 crc kubenswrapper[4696]: I0321 09:27:01.975222 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5_e4b15b92-b219-4ad6-a9ab-7d05a7130c92/pull/0.log" Mar 21 09:27:01 crc kubenswrapper[4696]: I0321 09:27:01.975476 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5_e4b15b92-b219-4ad6-a9ab-7d05a7130c92/pull/0.log" Mar 21 09:27:02 crc kubenswrapper[4696]: I0321 09:27:02.197687 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5_e4b15b92-b219-4ad6-a9ab-7d05a7130c92/util/0.log" Mar 21 09:27:02 crc kubenswrapper[4696]: I0321 09:27:02.250404 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5_e4b15b92-b219-4ad6-a9ab-7d05a7130c92/pull/0.log" Mar 21 09:27:02 crc kubenswrapper[4696]: I0321 09:27:02.377045 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5_e4b15b92-b219-4ad6-a9ab-7d05a7130c92/extract/0.log" Mar 21 09:27:02 crc kubenswrapper[4696]: I0321 09:27:02.472720 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh_8c8ec8ba-d404-4dca-aac7-3619ef7dbd75/util/0.log" Mar 21 09:27:02 crc kubenswrapper[4696]: I0321 09:27:02.739179 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh_8c8ec8ba-d404-4dca-aac7-3619ef7dbd75/pull/0.log" Mar 21 09:27:02 crc kubenswrapper[4696]: I0321 09:27:02.849341 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh_8c8ec8ba-d404-4dca-aac7-3619ef7dbd75/pull/0.log" Mar 21 09:27:02 crc kubenswrapper[4696]: I0321 09:27:02.933959 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh_8c8ec8ba-d404-4dca-aac7-3619ef7dbd75/util/0.log" Mar 21 09:27:02 crc kubenswrapper[4696]: I0321 09:27:02.982334 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh_8c8ec8ba-d404-4dca-aac7-3619ef7dbd75/pull/0.log" Mar 21 09:27:03 crc kubenswrapper[4696]: I0321 09:27:03.053660 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh_8c8ec8ba-d404-4dca-aac7-3619ef7dbd75/util/0.log" Mar 21 09:27:03 crc kubenswrapper[4696]: I0321 09:27:03.137009 4696 scope.go:117] "RemoveContainer" containerID="0835b4d56b56640ee07e7dda24d19c240cf0efb8d3a9ff0b21bc3ea9e413778e" Mar 21 09:27:03 crc kubenswrapper[4696]: I0321 09:27:03.251470 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh_8c8ec8ba-d404-4dca-aac7-3619ef7dbd75/extract/0.log" Mar 21 09:27:03 crc kubenswrapper[4696]: I0321 09:27:03.264317 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5vlg_e175ab46-d3f4-4051-a86e-dcf2727b9efd/extract-utilities/0.log" Mar 21 09:27:03 crc kubenswrapper[4696]: I0321 09:27:03.509267 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5vlg_e175ab46-d3f4-4051-a86e-dcf2727b9efd/extract-utilities/0.log" Mar 21 09:27:03 crc kubenswrapper[4696]: I0321 09:27:03.542863 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5vlg_e175ab46-d3f4-4051-a86e-dcf2727b9efd/extract-content/0.log" Mar 21 09:27:03 crc kubenswrapper[4696]: I0321 09:27:03.620289 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5vlg_e175ab46-d3f4-4051-a86e-dcf2727b9efd/extract-content/0.log" Mar 21 09:27:03 crc kubenswrapper[4696]: I0321 09:27:03.922472 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5vlg_e175ab46-d3f4-4051-a86e-dcf2727b9efd/extract-utilities/0.log" Mar 21 09:27:04 crc kubenswrapper[4696]: I0321 09:27:04.073201 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5vlg_e175ab46-d3f4-4051-a86e-dcf2727b9efd/extract-content/0.log" Mar 21 09:27:04 crc kubenswrapper[4696]: I0321 09:27:04.223629 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rwzg5_1050b6cc-b45d-42cb-8d54-7104988c8225/extract-utilities/0.log" Mar 21 09:27:04 crc kubenswrapper[4696]: I0321 09:27:04.602724 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rwzg5_1050b6cc-b45d-42cb-8d54-7104988c8225/extract-content/0.log" Mar 21 09:27:04 crc kubenswrapper[4696]: I0321 09:27:04.621853 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5vlg_e175ab46-d3f4-4051-a86e-dcf2727b9efd/registry-server/0.log" Mar 21 09:27:04 crc kubenswrapper[4696]: I0321 09:27:04.648958 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rwzg5_1050b6cc-b45d-42cb-8d54-7104988c8225/extract-content/0.log" Mar 21 09:27:04 crc kubenswrapper[4696]: I0321 09:27:04.771765 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rwzg5_1050b6cc-b45d-42cb-8d54-7104988c8225/extract-utilities/0.log" Mar 21 09:27:04 crc kubenswrapper[4696]: I0321 09:27:04.912961 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rwzg5_1050b6cc-b45d-42cb-8d54-7104988c8225/extract-utilities/0.log" Mar 21 09:27:05 crc kubenswrapper[4696]: I0321 09:27:05.010872 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rwzg5_1050b6cc-b45d-42cb-8d54-7104988c8225/extract-content/0.log" Mar 21 09:27:05 crc kubenswrapper[4696]: I0321 09:27:05.302066 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-w6xbt_d45283e8-56c7-4a2e-8a7e-88e459bb1c04/marketplace-operator/0.log" Mar 21 09:27:05 crc kubenswrapper[4696]: I0321 09:27:05.575277 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mmkz6_5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77/extract-utilities/0.log" Mar 21 09:27:05 crc kubenswrapper[4696]: I0321 09:27:05.633942 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rwzg5_1050b6cc-b45d-42cb-8d54-7104988c8225/registry-server/0.log" Mar 21 09:27:05 crc kubenswrapper[4696]: I0321 09:27:05.697738 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mmkz6_5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77/extract-content/0.log" Mar 21 09:27:05 crc kubenswrapper[4696]: I0321 09:27:05.807147 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mmkz6_5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77/extract-utilities/0.log" Mar 21 09:27:05 crc kubenswrapper[4696]: I0321 09:27:05.823621 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mmkz6_5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77/extract-content/0.log" Mar 21 09:27:06 crc kubenswrapper[4696]: I0321 09:27:06.165231 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mmkz6_5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77/extract-utilities/0.log" Mar 21 09:27:06 crc kubenswrapper[4696]: I0321 09:27:06.253261 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mmkz6_5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77/registry-server/0.log" Mar 21 09:27:06 crc kubenswrapper[4696]: I0321 09:27:06.296595 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mmkz6_5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77/extract-content/0.log" Mar 21 09:27:06 crc kubenswrapper[4696]: I0321 09:27:06.307352 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6f4n_b49f2cd9-95cf-4749-889e-c47d027c253b/extract-utilities/0.log" Mar 21 09:27:06 crc kubenswrapper[4696]: I0321 09:27:06.523644 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6f4n_b49f2cd9-95cf-4749-889e-c47d027c253b/extract-utilities/0.log" Mar 21 09:27:06 crc kubenswrapper[4696]: I0321 09:27:06.545976 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6f4n_b49f2cd9-95cf-4749-889e-c47d027c253b/extract-content/0.log" Mar 21 09:27:06 crc kubenswrapper[4696]: I0321 09:27:06.592189 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6f4n_b49f2cd9-95cf-4749-889e-c47d027c253b/extract-content/0.log" Mar 21 09:27:06 crc kubenswrapper[4696]: I0321 09:27:06.836177 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6f4n_b49f2cd9-95cf-4749-889e-c47d027c253b/extract-content/0.log" Mar 21 09:27:06 crc kubenswrapper[4696]: I0321 09:27:06.874912 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6f4n_b49f2cd9-95cf-4749-889e-c47d027c253b/extract-utilities/0.log" Mar 21 09:27:07 crc kubenswrapper[4696]: I0321 09:27:07.388696 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6f4n_b49f2cd9-95cf-4749-889e-c47d027c253b/registry-server/0.log" Mar 21 09:27:27 crc kubenswrapper[4696]: I0321 09:27:27.674371 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-db48b98cb-l8q89_d99a53b9-d503-4410-9c7c-807fca5276bc/prometheus-operator-admission-webhook/0.log" Mar 21 09:27:27 crc kubenswrapper[4696]: I0321 09:27:27.788012 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-8ff7d675-rs27t_527ae575-3dfd-4f0a-9af9-63668af676d3/prometheus-operator/0.log" Mar 21 09:27:27 crc kubenswrapper[4696]: I0321 09:27:27.893513 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-db48b98cb-78qsn_474e7a66-fb1d-448c-8297-b511ec7b6e7c/prometheus-operator-admission-webhook/0.log" Mar 21 09:27:28 crc kubenswrapper[4696]: I0321 09:27:28.096045 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-64bcb4cf6f-jjk5b_127ee197-aa20-41d3-ac26-e8aead566b64/perses-operator/0.log" Mar 21 09:27:28 crc kubenswrapper[4696]: I0321 09:27:28.126068 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-6dd7dd855f-sp4gg_e864a9cc-e270-4238-841b-77d0abad085d/operator/0.log" Mar 21 09:27:48 crc kubenswrapper[4696]: I0321 09:27:48.559893 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5685547f78-5x2wx_75d1fdab-70ac-46df-8b60-3e2f75985686/manager/0.log" Mar 21 09:27:48 crc kubenswrapper[4696]: I0321 09:27:48.985292 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5685547f78-5x2wx_75d1fdab-70ac-46df-8b60-3e2f75985686/kube-rbac-proxy/0.log" Mar 21 09:28:00 crc kubenswrapper[4696]: I0321 09:28:00.176586 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568088-w2ckn"] Mar 21 09:28:00 crc kubenswrapper[4696]: E0321 09:28:00.177914 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57992ac0-359d-43ba-b9f3-d5e50e96cebd" containerName="oc" Mar 21 09:28:00 crc kubenswrapper[4696]: I0321 09:28:00.177929 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="57992ac0-359d-43ba-b9f3-d5e50e96cebd" containerName="oc" Mar 21 09:28:00 crc kubenswrapper[4696]: I0321 09:28:00.178376 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="57992ac0-359d-43ba-b9f3-d5e50e96cebd" containerName="oc" Mar 21 09:28:00 crc kubenswrapper[4696]: I0321 09:28:00.179374 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568088-w2ckn" Mar 21 09:28:00 crc kubenswrapper[4696]: I0321 09:28:00.195608 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:28:00 crc kubenswrapper[4696]: I0321 09:28:00.196052 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:28:00 crc kubenswrapper[4696]: I0321 09:28:00.196241 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:28:00 crc kubenswrapper[4696]: I0321 09:28:00.292274 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568088-w2ckn"] Mar 21 09:28:00 crc kubenswrapper[4696]: I0321 09:28:00.341568 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtdtg\" (UniqueName: \"kubernetes.io/projected/b3a7dacd-a824-4d86-bf9b-57c1cfa220df-kube-api-access-vtdtg\") pod \"auto-csr-approver-29568088-w2ckn\" (UID: \"b3a7dacd-a824-4d86-bf9b-57c1cfa220df\") " pod="openshift-infra/auto-csr-approver-29568088-w2ckn" Mar 21 09:28:00 crc kubenswrapper[4696]: I0321 09:28:00.444161 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtdtg\" (UniqueName: \"kubernetes.io/projected/b3a7dacd-a824-4d86-bf9b-57c1cfa220df-kube-api-access-vtdtg\") pod \"auto-csr-approver-29568088-w2ckn\" (UID: \"b3a7dacd-a824-4d86-bf9b-57c1cfa220df\") " pod="openshift-infra/auto-csr-approver-29568088-w2ckn" Mar 21 09:28:00 crc kubenswrapper[4696]: I0321 09:28:00.474883 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtdtg\" (UniqueName: \"kubernetes.io/projected/b3a7dacd-a824-4d86-bf9b-57c1cfa220df-kube-api-access-vtdtg\") pod \"auto-csr-approver-29568088-w2ckn\" (UID: \"b3a7dacd-a824-4d86-bf9b-57c1cfa220df\") " pod="openshift-infra/auto-csr-approver-29568088-w2ckn" Mar 21 09:28:00 crc kubenswrapper[4696]: I0321 09:28:00.547651 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568088-w2ckn" Mar 21 09:28:01 crc kubenswrapper[4696]: I0321 09:28:01.325939 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568088-w2ckn"] Mar 21 09:28:01 crc kubenswrapper[4696]: I0321 09:28:01.934273 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568088-w2ckn" event={"ID":"b3a7dacd-a824-4d86-bf9b-57c1cfa220df","Type":"ContainerStarted","Data":"ebe6ac585a1294b985f8d9ca359c3f0ed363a0485ba2dcef9e0446c069f34776"} Mar 21 09:28:02 crc kubenswrapper[4696]: I0321 09:28:02.946950 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568088-w2ckn" event={"ID":"b3a7dacd-a824-4d86-bf9b-57c1cfa220df","Type":"ContainerStarted","Data":"19e31703b8381b0f31d344e849892b111ffc7efa6e5589d7a48a416d6db45179"} Mar 21 09:28:02 crc kubenswrapper[4696]: I0321 09:28:02.973305 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29568088-w2ckn" podStartSLOduration=1.834078092 podStartE2EDuration="2.973286882s" podCreationTimestamp="2026-03-21 09:28:00 +0000 UTC" firstStartedPulling="2026-03-21 09:28:01.338074805 +0000 UTC m=+3615.458955508" lastFinishedPulling="2026-03-21 09:28:02.477283585 +0000 UTC m=+3616.598164298" observedRunningTime="2026-03-21 09:28:02.96212368 +0000 UTC m=+3617.083004393" watchObservedRunningTime="2026-03-21 09:28:02.973286882 +0000 UTC m=+3617.094167585" Mar 21 09:28:03 crc kubenswrapper[4696]: I0321 09:28:03.955357 4696 generic.go:334] "Generic (PLEG): container finished" podID="b3a7dacd-a824-4d86-bf9b-57c1cfa220df" containerID="19e31703b8381b0f31d344e849892b111ffc7efa6e5589d7a48a416d6db45179" exitCode=0 Mar 21 09:28:03 crc kubenswrapper[4696]: I0321 09:28:03.955634 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568088-w2ckn" event={"ID":"b3a7dacd-a824-4d86-bf9b-57c1cfa220df","Type":"ContainerDied","Data":"19e31703b8381b0f31d344e849892b111ffc7efa6e5589d7a48a416d6db45179"} Mar 21 09:28:06 crc kubenswrapper[4696]: I0321 09:28:06.149663 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568088-w2ckn" Mar 21 09:28:06 crc kubenswrapper[4696]: I0321 09:28:06.314007 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtdtg\" (UniqueName: \"kubernetes.io/projected/b3a7dacd-a824-4d86-bf9b-57c1cfa220df-kube-api-access-vtdtg\") pod \"b3a7dacd-a824-4d86-bf9b-57c1cfa220df\" (UID: \"b3a7dacd-a824-4d86-bf9b-57c1cfa220df\") " Mar 21 09:28:06 crc kubenswrapper[4696]: I0321 09:28:06.331622 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3a7dacd-a824-4d86-bf9b-57c1cfa220df-kube-api-access-vtdtg" (OuterVolumeSpecName: "kube-api-access-vtdtg") pod "b3a7dacd-a824-4d86-bf9b-57c1cfa220df" (UID: "b3a7dacd-a824-4d86-bf9b-57c1cfa220df"). InnerVolumeSpecName "kube-api-access-vtdtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:28:06 crc kubenswrapper[4696]: I0321 09:28:06.419409 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtdtg\" (UniqueName: \"kubernetes.io/projected/b3a7dacd-a824-4d86-bf9b-57c1cfa220df-kube-api-access-vtdtg\") on node \"crc\" DevicePath \"\"" Mar 21 09:28:06 crc kubenswrapper[4696]: I0321 09:28:06.991841 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568088-w2ckn" event={"ID":"b3a7dacd-a824-4d86-bf9b-57c1cfa220df","Type":"ContainerDied","Data":"ebe6ac585a1294b985f8d9ca359c3f0ed363a0485ba2dcef9e0446c069f34776"} Mar 21 09:28:06 crc kubenswrapper[4696]: I0321 09:28:06.991877 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebe6ac585a1294b985f8d9ca359c3f0ed363a0485ba2dcef9e0446c069f34776" Mar 21 09:28:06 crc kubenswrapper[4696]: I0321 09:28:06.991919 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568088-w2ckn" Mar 21 09:28:07 crc kubenswrapper[4696]: I0321 09:28:07.222122 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568082-s2tkt"] Mar 21 09:28:07 crc kubenswrapper[4696]: I0321 09:28:07.231173 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568082-s2tkt"] Mar 21 09:28:08 crc kubenswrapper[4696]: I0321 09:28:08.546014 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60fffaff-5238-4b84-882b-16a0b4c542f6" path="/var/lib/kubelet/pods/60fffaff-5238-4b84-882b-16a0b4c542f6/volumes" Mar 21 09:28:30 crc kubenswrapper[4696]: I0321 09:28:30.341088 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:28:30 crc kubenswrapper[4696]: I0321 09:28:30.341616 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:29:00 crc kubenswrapper[4696]: I0321 09:29:00.341082 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:29:00 crc kubenswrapper[4696]: I0321 09:29:00.341610 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:29:03 crc kubenswrapper[4696]: I0321 09:29:03.255479 4696 scope.go:117] "RemoveContainer" containerID="9e183804d43fe19470aed23665c5decfa6392ab09c92190bb8893e5546519f82" Mar 21 09:29:30 crc kubenswrapper[4696]: I0321 09:29:30.341143 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:29:30 crc kubenswrapper[4696]: I0321 09:29:30.341683 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:29:30 crc kubenswrapper[4696]: I0321 09:29:30.341725 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 09:29:30 crc kubenswrapper[4696]: I0321 09:29:30.342579 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b"} pod="openshift-machine-config-operator/machine-config-daemon-z7srw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 09:29:30 crc kubenswrapper[4696]: I0321 09:29:30.342633 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" containerID="cri-o://32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" gracePeriod=600 Mar 21 09:29:30 crc kubenswrapper[4696]: E0321 09:29:30.473379 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:29:30 crc kubenswrapper[4696]: I0321 09:29:30.786357 4696 generic.go:334] "Generic (PLEG): container finished" podID="daaf227a-2305-495e-8495-a280abcd8e10" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" exitCode=0 Mar 21 09:29:30 crc kubenswrapper[4696]: I0321 09:29:30.786398 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerDied","Data":"32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b"} Mar 21 09:29:30 crc kubenswrapper[4696]: I0321 09:29:30.786429 4696 scope.go:117] "RemoveContainer" containerID="aa5a03c21bdd6334cc43af15b5fd004e41076b2d7b35f63b8f51c60a2bc0915a" Mar 21 09:29:30 crc kubenswrapper[4696]: I0321 09:29:30.787158 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:29:30 crc kubenswrapper[4696]: E0321 09:29:30.787542 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:29:42 crc kubenswrapper[4696]: I0321 09:29:42.541451 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:29:42 crc kubenswrapper[4696]: E0321 09:29:42.542118 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:29:56 crc kubenswrapper[4696]: I0321 09:29:56.546005 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:29:56 crc kubenswrapper[4696]: E0321 09:29:56.546732 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.152509 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568090-g4t47"] Mar 21 09:30:00 crc kubenswrapper[4696]: E0321 09:30:00.153516 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3a7dacd-a824-4d86-bf9b-57c1cfa220df" containerName="oc" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.153530 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3a7dacd-a824-4d86-bf9b-57c1cfa220df" containerName="oc" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.153716 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3a7dacd-a824-4d86-bf9b-57c1cfa220df" containerName="oc" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.154483 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568090-g4t47" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.156537 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.156920 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.157591 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.165535 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568090-s45qn"] Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.167216 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568090-s45qn" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.171616 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.171805 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.179395 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568090-g4t47"] Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.188461 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568090-s45qn"] Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.283016 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88w4c\" (UniqueName: \"kubernetes.io/projected/bb1746c4-d2da-45f7-90a2-d28635623d57-kube-api-access-88w4c\") pod \"auto-csr-approver-29568090-g4t47\" (UID: \"bb1746c4-d2da-45f7-90a2-d28635623d57\") " pod="openshift-infra/auto-csr-approver-29568090-g4t47" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.283106 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfdnz\" (UniqueName: \"kubernetes.io/projected/68000673-4e7e-4b9b-b45a-4c1172bc160b-kube-api-access-hfdnz\") pod \"collect-profiles-29568090-s45qn\" (UID: \"68000673-4e7e-4b9b-b45a-4c1172bc160b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568090-s45qn" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.283157 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/68000673-4e7e-4b9b-b45a-4c1172bc160b-secret-volume\") pod \"collect-profiles-29568090-s45qn\" (UID: \"68000673-4e7e-4b9b-b45a-4c1172bc160b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568090-s45qn" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.283375 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/68000673-4e7e-4b9b-b45a-4c1172bc160b-config-volume\") pod \"collect-profiles-29568090-s45qn\" (UID: \"68000673-4e7e-4b9b-b45a-4c1172bc160b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568090-s45qn" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.385615 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/68000673-4e7e-4b9b-b45a-4c1172bc160b-config-volume\") pod \"collect-profiles-29568090-s45qn\" (UID: \"68000673-4e7e-4b9b-b45a-4c1172bc160b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568090-s45qn" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.385702 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88w4c\" (UniqueName: \"kubernetes.io/projected/bb1746c4-d2da-45f7-90a2-d28635623d57-kube-api-access-88w4c\") pod \"auto-csr-approver-29568090-g4t47\" (UID: \"bb1746c4-d2da-45f7-90a2-d28635623d57\") " pod="openshift-infra/auto-csr-approver-29568090-g4t47" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.385745 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfdnz\" (UniqueName: \"kubernetes.io/projected/68000673-4e7e-4b9b-b45a-4c1172bc160b-kube-api-access-hfdnz\") pod \"collect-profiles-29568090-s45qn\" (UID: \"68000673-4e7e-4b9b-b45a-4c1172bc160b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568090-s45qn" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.385797 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/68000673-4e7e-4b9b-b45a-4c1172bc160b-secret-volume\") pod \"collect-profiles-29568090-s45qn\" (UID: \"68000673-4e7e-4b9b-b45a-4c1172bc160b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568090-s45qn" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.387046 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/68000673-4e7e-4b9b-b45a-4c1172bc160b-config-volume\") pod \"collect-profiles-29568090-s45qn\" (UID: \"68000673-4e7e-4b9b-b45a-4c1172bc160b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568090-s45qn" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.392796 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/68000673-4e7e-4b9b-b45a-4c1172bc160b-secret-volume\") pod \"collect-profiles-29568090-s45qn\" (UID: \"68000673-4e7e-4b9b-b45a-4c1172bc160b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568090-s45qn" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.405690 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfdnz\" (UniqueName: \"kubernetes.io/projected/68000673-4e7e-4b9b-b45a-4c1172bc160b-kube-api-access-hfdnz\") pod \"collect-profiles-29568090-s45qn\" (UID: \"68000673-4e7e-4b9b-b45a-4c1172bc160b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568090-s45qn" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.412997 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88w4c\" (UniqueName: \"kubernetes.io/projected/bb1746c4-d2da-45f7-90a2-d28635623d57-kube-api-access-88w4c\") pod \"auto-csr-approver-29568090-g4t47\" (UID: \"bb1746c4-d2da-45f7-90a2-d28635623d57\") " pod="openshift-infra/auto-csr-approver-29568090-g4t47" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.481663 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568090-g4t47" Mar 21 09:30:00 crc kubenswrapper[4696]: I0321 09:30:00.503846 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568090-s45qn" Mar 21 09:30:01 crc kubenswrapper[4696]: I0321 09:30:01.367054 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568090-g4t47"] Mar 21 09:30:01 crc kubenswrapper[4696]: I0321 09:30:01.390597 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 09:30:01 crc kubenswrapper[4696]: W0321 09:30:01.725375 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68000673_4e7e_4b9b_b45a_4c1172bc160b.slice/crio-4adc59e84570ef0a123835d9c19292f942321f7b9a87b54687a0ec094cb7589b WatchSource:0}: Error finding container 4adc59e84570ef0a123835d9c19292f942321f7b9a87b54687a0ec094cb7589b: Status 404 returned error can't find the container with id 4adc59e84570ef0a123835d9c19292f942321f7b9a87b54687a0ec094cb7589b Mar 21 09:30:01 crc kubenswrapper[4696]: I0321 09:30:01.730892 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568090-s45qn"] Mar 21 09:30:02 crc kubenswrapper[4696]: I0321 09:30:02.082902 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568090-g4t47" event={"ID":"bb1746c4-d2da-45f7-90a2-d28635623d57","Type":"ContainerStarted","Data":"c76a2fbe60877a946dddbace7e3f07fe1fd94d0e9d39922f0edf4b05331638f5"} Mar 21 09:30:02 crc kubenswrapper[4696]: I0321 09:30:02.084687 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568090-s45qn" event={"ID":"68000673-4e7e-4b9b-b45a-4c1172bc160b","Type":"ContainerStarted","Data":"254e45d032146128e8f081c60d05da88f57256f766eeadd095fdd44a259c04b3"} Mar 21 09:30:02 crc kubenswrapper[4696]: I0321 09:30:02.084743 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568090-s45qn" event={"ID":"68000673-4e7e-4b9b-b45a-4c1172bc160b","Type":"ContainerStarted","Data":"4adc59e84570ef0a123835d9c19292f942321f7b9a87b54687a0ec094cb7589b"} Mar 21 09:30:02 crc kubenswrapper[4696]: I0321 09:30:02.115004 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29568090-s45qn" podStartSLOduration=2.114980886 podStartE2EDuration="2.114980886s" podCreationTimestamp="2026-03-21 09:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 09:30:02.102766347 +0000 UTC m=+3736.223647070" watchObservedRunningTime="2026-03-21 09:30:02.114980886 +0000 UTC m=+3736.235861599" Mar 21 09:30:03 crc kubenswrapper[4696]: I0321 09:30:03.094607 4696 generic.go:334] "Generic (PLEG): container finished" podID="68000673-4e7e-4b9b-b45a-4c1172bc160b" containerID="254e45d032146128e8f081c60d05da88f57256f766eeadd095fdd44a259c04b3" exitCode=0 Mar 21 09:30:03 crc kubenswrapper[4696]: I0321 09:30:03.094657 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568090-s45qn" event={"ID":"68000673-4e7e-4b9b-b45a-4c1172bc160b","Type":"ContainerDied","Data":"254e45d032146128e8f081c60d05da88f57256f766eeadd095fdd44a259c04b3"} Mar 21 09:30:04 crc kubenswrapper[4696]: I0321 09:30:04.106513 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568090-g4t47" event={"ID":"bb1746c4-d2da-45f7-90a2-d28635623d57","Type":"ContainerStarted","Data":"3884babbc72e0c022df25b381e932e30daa732179936a1c41eec0c731632e3d1"} Mar 21 09:30:04 crc kubenswrapper[4696]: I0321 09:30:04.123213 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29568090-g4t47" podStartSLOduration=2.069800654 podStartE2EDuration="4.123196034s" podCreationTimestamp="2026-03-21 09:30:00 +0000 UTC" firstStartedPulling="2026-03-21 09:30:01.390366809 +0000 UTC m=+3735.511247522" lastFinishedPulling="2026-03-21 09:30:03.443762189 +0000 UTC m=+3737.564642902" observedRunningTime="2026-03-21 09:30:04.121998489 +0000 UTC m=+3738.242879202" watchObservedRunningTime="2026-03-21 09:30:04.123196034 +0000 UTC m=+3738.244076747" Mar 21 09:30:05 crc kubenswrapper[4696]: I0321 09:30:05.116158 4696 generic.go:334] "Generic (PLEG): container finished" podID="bb1746c4-d2da-45f7-90a2-d28635623d57" containerID="3884babbc72e0c022df25b381e932e30daa732179936a1c41eec0c731632e3d1" exitCode=0 Mar 21 09:30:05 crc kubenswrapper[4696]: I0321 09:30:05.116390 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568090-g4t47" event={"ID":"bb1746c4-d2da-45f7-90a2-d28635623d57","Type":"ContainerDied","Data":"3884babbc72e0c022df25b381e932e30daa732179936a1c41eec0c731632e3d1"} Mar 21 09:30:05 crc kubenswrapper[4696]: I0321 09:30:05.328493 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568090-s45qn" Mar 21 09:30:05 crc kubenswrapper[4696]: I0321 09:30:05.406495 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfdnz\" (UniqueName: \"kubernetes.io/projected/68000673-4e7e-4b9b-b45a-4c1172bc160b-kube-api-access-hfdnz\") pod \"68000673-4e7e-4b9b-b45a-4c1172bc160b\" (UID: \"68000673-4e7e-4b9b-b45a-4c1172bc160b\") " Mar 21 09:30:05 crc kubenswrapper[4696]: I0321 09:30:05.407644 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/68000673-4e7e-4b9b-b45a-4c1172bc160b-config-volume\") pod \"68000673-4e7e-4b9b-b45a-4c1172bc160b\" (UID: \"68000673-4e7e-4b9b-b45a-4c1172bc160b\") " Mar 21 09:30:05 crc kubenswrapper[4696]: I0321 09:30:05.407801 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/68000673-4e7e-4b9b-b45a-4c1172bc160b-secret-volume\") pod \"68000673-4e7e-4b9b-b45a-4c1172bc160b\" (UID: \"68000673-4e7e-4b9b-b45a-4c1172bc160b\") " Mar 21 09:30:05 crc kubenswrapper[4696]: I0321 09:30:05.408199 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68000673-4e7e-4b9b-b45a-4c1172bc160b-config-volume" (OuterVolumeSpecName: "config-volume") pod "68000673-4e7e-4b9b-b45a-4c1172bc160b" (UID: "68000673-4e7e-4b9b-b45a-4c1172bc160b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 09:30:05 crc kubenswrapper[4696]: I0321 09:30:05.408918 4696 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/68000673-4e7e-4b9b-b45a-4c1172bc160b-config-volume\") on node \"crc\" DevicePath \"\"" Mar 21 09:30:05 crc kubenswrapper[4696]: I0321 09:30:05.412081 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68000673-4e7e-4b9b-b45a-4c1172bc160b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "68000673-4e7e-4b9b-b45a-4c1172bc160b" (UID: "68000673-4e7e-4b9b-b45a-4c1172bc160b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:30:05 crc kubenswrapper[4696]: I0321 09:30:05.417615 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68000673-4e7e-4b9b-b45a-4c1172bc160b-kube-api-access-hfdnz" (OuterVolumeSpecName: "kube-api-access-hfdnz") pod "68000673-4e7e-4b9b-b45a-4c1172bc160b" (UID: "68000673-4e7e-4b9b-b45a-4c1172bc160b"). InnerVolumeSpecName "kube-api-access-hfdnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:30:05 crc kubenswrapper[4696]: I0321 09:30:05.511164 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfdnz\" (UniqueName: \"kubernetes.io/projected/68000673-4e7e-4b9b-b45a-4c1172bc160b-kube-api-access-hfdnz\") on node \"crc\" DevicePath \"\"" Mar 21 09:30:05 crc kubenswrapper[4696]: I0321 09:30:05.511200 4696 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/68000673-4e7e-4b9b-b45a-4c1172bc160b-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 21 09:30:06 crc kubenswrapper[4696]: I0321 09:30:06.127067 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568090-s45qn" event={"ID":"68000673-4e7e-4b9b-b45a-4c1172bc160b","Type":"ContainerDied","Data":"4adc59e84570ef0a123835d9c19292f942321f7b9a87b54687a0ec094cb7589b"} Mar 21 09:30:06 crc kubenswrapper[4696]: I0321 09:30:06.127134 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4adc59e84570ef0a123835d9c19292f942321f7b9a87b54687a0ec094cb7589b" Mar 21 09:30:06 crc kubenswrapper[4696]: I0321 09:30:06.127085 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568090-s45qn" Mar 21 09:30:06 crc kubenswrapper[4696]: I0321 09:30:06.416707 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568045-xllgs"] Mar 21 09:30:06 crc kubenswrapper[4696]: I0321 09:30:06.426028 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568045-xllgs"] Mar 21 09:30:06 crc kubenswrapper[4696]: I0321 09:30:06.561920 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="896b4905-5c3e-4fff-8c28-3891ae88049d" path="/var/lib/kubelet/pods/896b4905-5c3e-4fff-8c28-3891ae88049d/volumes" Mar 21 09:30:07 crc kubenswrapper[4696]: I0321 09:30:07.136759 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568090-g4t47" event={"ID":"bb1746c4-d2da-45f7-90a2-d28635623d57","Type":"ContainerDied","Data":"c76a2fbe60877a946dddbace7e3f07fe1fd94d0e9d39922f0edf4b05331638f5"} Mar 21 09:30:07 crc kubenswrapper[4696]: I0321 09:30:07.137104 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c76a2fbe60877a946dddbace7e3f07fe1fd94d0e9d39922f0edf4b05331638f5" Mar 21 09:30:07 crc kubenswrapper[4696]: I0321 09:30:07.187611 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568090-g4t47" Mar 21 09:30:07 crc kubenswrapper[4696]: I0321 09:30:07.263307 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88w4c\" (UniqueName: \"kubernetes.io/projected/bb1746c4-d2da-45f7-90a2-d28635623d57-kube-api-access-88w4c\") pod \"bb1746c4-d2da-45f7-90a2-d28635623d57\" (UID: \"bb1746c4-d2da-45f7-90a2-d28635623d57\") " Mar 21 09:30:07 crc kubenswrapper[4696]: I0321 09:30:07.279610 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb1746c4-d2da-45f7-90a2-d28635623d57-kube-api-access-88w4c" (OuterVolumeSpecName: "kube-api-access-88w4c") pod "bb1746c4-d2da-45f7-90a2-d28635623d57" (UID: "bb1746c4-d2da-45f7-90a2-d28635623d57"). InnerVolumeSpecName "kube-api-access-88w4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:30:07 crc kubenswrapper[4696]: I0321 09:30:07.366400 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88w4c\" (UniqueName: \"kubernetes.io/projected/bb1746c4-d2da-45f7-90a2-d28635623d57-kube-api-access-88w4c\") on node \"crc\" DevicePath \"\"" Mar 21 09:30:08 crc kubenswrapper[4696]: I0321 09:30:08.146907 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568090-g4t47" Mar 21 09:30:08 crc kubenswrapper[4696]: I0321 09:30:08.250172 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568084-b55mn"] Mar 21 09:30:08 crc kubenswrapper[4696]: I0321 09:30:08.258655 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568084-b55mn"] Mar 21 09:30:08 crc kubenswrapper[4696]: I0321 09:30:08.535087 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:30:08 crc kubenswrapper[4696]: E0321 09:30:08.535417 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:30:08 crc kubenswrapper[4696]: I0321 09:30:08.547853 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff83029d-26f8-46b7-b54f-5a8fd3b8598e" path="/var/lib/kubelet/pods/ff83029d-26f8-46b7-b54f-5a8fd3b8598e/volumes" Mar 21 09:30:22 crc kubenswrapper[4696]: I0321 09:30:22.535246 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:30:22 crc kubenswrapper[4696]: E0321 09:30:22.535945 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:30:28 crc kubenswrapper[4696]: I0321 09:30:28.318672 4696 generic.go:334] "Generic (PLEG): container finished" podID="6d6e8c8f-6a59-443a-9874-848be975ebd1" containerID="a1511802f4c23a501730dbcfa0f1b4f0babb12187a6dd728849dc817af142b12" exitCode=0 Mar 21 09:30:28 crc kubenswrapper[4696]: I0321 09:30:28.318753 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mckgv/must-gather-kqwf9" event={"ID":"6d6e8c8f-6a59-443a-9874-848be975ebd1","Type":"ContainerDied","Data":"a1511802f4c23a501730dbcfa0f1b4f0babb12187a6dd728849dc817af142b12"} Mar 21 09:30:28 crc kubenswrapper[4696]: I0321 09:30:28.319996 4696 scope.go:117] "RemoveContainer" containerID="a1511802f4c23a501730dbcfa0f1b4f0babb12187a6dd728849dc817af142b12" Mar 21 09:30:29 crc kubenswrapper[4696]: I0321 09:30:29.101553 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mckgv_must-gather-kqwf9_6d6e8c8f-6a59-443a-9874-848be975ebd1/gather/0.log" Mar 21 09:30:35 crc kubenswrapper[4696]: I0321 09:30:35.535976 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:30:35 crc kubenswrapper[4696]: E0321 09:30:35.536908 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:30:37 crc kubenswrapper[4696]: I0321 09:30:37.356873 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mckgv/must-gather-kqwf9"] Mar 21 09:30:37 crc kubenswrapper[4696]: I0321 09:30:37.358080 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-mckgv/must-gather-kqwf9" podUID="6d6e8c8f-6a59-443a-9874-848be975ebd1" containerName="copy" containerID="cri-o://846c1d1531ac135a588967c945e79fda57f16e40397e1a85b1155ffd22a198e2" gracePeriod=2 Mar 21 09:30:37 crc kubenswrapper[4696]: I0321 09:30:37.367526 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mckgv/must-gather-kqwf9"] Mar 21 09:30:38 crc kubenswrapper[4696]: I0321 09:30:38.343113 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mckgv_must-gather-kqwf9_6d6e8c8f-6a59-443a-9874-848be975ebd1/copy/0.log" Mar 21 09:30:38 crc kubenswrapper[4696]: I0321 09:30:38.344097 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mckgv/must-gather-kqwf9" Mar 21 09:30:38 crc kubenswrapper[4696]: I0321 09:30:38.448416 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mckgv_must-gather-kqwf9_6d6e8c8f-6a59-443a-9874-848be975ebd1/copy/0.log" Mar 21 09:30:38 crc kubenswrapper[4696]: I0321 09:30:38.450254 4696 generic.go:334] "Generic (PLEG): container finished" podID="6d6e8c8f-6a59-443a-9874-848be975ebd1" containerID="846c1d1531ac135a588967c945e79fda57f16e40397e1a85b1155ffd22a198e2" exitCode=143 Mar 21 09:30:38 crc kubenswrapper[4696]: I0321 09:30:38.450313 4696 scope.go:117] "RemoveContainer" containerID="846c1d1531ac135a588967c945e79fda57f16e40397e1a85b1155ffd22a198e2" Mar 21 09:30:38 crc kubenswrapper[4696]: I0321 09:30:38.450332 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mckgv/must-gather-kqwf9" Mar 21 09:30:38 crc kubenswrapper[4696]: I0321 09:30:38.476291 4696 scope.go:117] "RemoveContainer" containerID="a1511802f4c23a501730dbcfa0f1b4f0babb12187a6dd728849dc817af142b12" Mar 21 09:30:38 crc kubenswrapper[4696]: I0321 09:30:38.513092 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6d6e8c8f-6a59-443a-9874-848be975ebd1-must-gather-output\") pod \"6d6e8c8f-6a59-443a-9874-848be975ebd1\" (UID: \"6d6e8c8f-6a59-443a-9874-848be975ebd1\") " Mar 21 09:30:38 crc kubenswrapper[4696]: I0321 09:30:38.513339 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7wp6\" (UniqueName: \"kubernetes.io/projected/6d6e8c8f-6a59-443a-9874-848be975ebd1-kube-api-access-q7wp6\") pod \"6d6e8c8f-6a59-443a-9874-848be975ebd1\" (UID: \"6d6e8c8f-6a59-443a-9874-848be975ebd1\") " Mar 21 09:30:38 crc kubenswrapper[4696]: I0321 09:30:38.519886 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d6e8c8f-6a59-443a-9874-848be975ebd1-kube-api-access-q7wp6" (OuterVolumeSpecName: "kube-api-access-q7wp6") pod "6d6e8c8f-6a59-443a-9874-848be975ebd1" (UID: "6d6e8c8f-6a59-443a-9874-848be975ebd1"). InnerVolumeSpecName "kube-api-access-q7wp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:30:38 crc kubenswrapper[4696]: I0321 09:30:38.604739 4696 scope.go:117] "RemoveContainer" containerID="846c1d1531ac135a588967c945e79fda57f16e40397e1a85b1155ffd22a198e2" Mar 21 09:30:38 crc kubenswrapper[4696]: E0321 09:30:38.606543 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"846c1d1531ac135a588967c945e79fda57f16e40397e1a85b1155ffd22a198e2\": container with ID starting with 846c1d1531ac135a588967c945e79fda57f16e40397e1a85b1155ffd22a198e2 not found: ID does not exist" containerID="846c1d1531ac135a588967c945e79fda57f16e40397e1a85b1155ffd22a198e2" Mar 21 09:30:38 crc kubenswrapper[4696]: I0321 09:30:38.606648 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"846c1d1531ac135a588967c945e79fda57f16e40397e1a85b1155ffd22a198e2"} err="failed to get container status \"846c1d1531ac135a588967c945e79fda57f16e40397e1a85b1155ffd22a198e2\": rpc error: code = NotFound desc = could not find container \"846c1d1531ac135a588967c945e79fda57f16e40397e1a85b1155ffd22a198e2\": container with ID starting with 846c1d1531ac135a588967c945e79fda57f16e40397e1a85b1155ffd22a198e2 not found: ID does not exist" Mar 21 09:30:38 crc kubenswrapper[4696]: I0321 09:30:38.606747 4696 scope.go:117] "RemoveContainer" containerID="a1511802f4c23a501730dbcfa0f1b4f0babb12187a6dd728849dc817af142b12" Mar 21 09:30:38 crc kubenswrapper[4696]: E0321 09:30:38.607689 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1511802f4c23a501730dbcfa0f1b4f0babb12187a6dd728849dc817af142b12\": container with ID starting with a1511802f4c23a501730dbcfa0f1b4f0babb12187a6dd728849dc817af142b12 not found: ID does not exist" containerID="a1511802f4c23a501730dbcfa0f1b4f0babb12187a6dd728849dc817af142b12" Mar 21 09:30:38 crc kubenswrapper[4696]: I0321 09:30:38.607732 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1511802f4c23a501730dbcfa0f1b4f0babb12187a6dd728849dc817af142b12"} err="failed to get container status \"a1511802f4c23a501730dbcfa0f1b4f0babb12187a6dd728849dc817af142b12\": rpc error: code = NotFound desc = could not find container \"a1511802f4c23a501730dbcfa0f1b4f0babb12187a6dd728849dc817af142b12\": container with ID starting with a1511802f4c23a501730dbcfa0f1b4f0babb12187a6dd728849dc817af142b12 not found: ID does not exist" Mar 21 09:30:38 crc kubenswrapper[4696]: I0321 09:30:38.616083 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7wp6\" (UniqueName: \"kubernetes.io/projected/6d6e8c8f-6a59-443a-9874-848be975ebd1-kube-api-access-q7wp6\") on node \"crc\" DevicePath \"\"" Mar 21 09:30:38 crc kubenswrapper[4696]: I0321 09:30:38.719133 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d6e8c8f-6a59-443a-9874-848be975ebd1-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "6d6e8c8f-6a59-443a-9874-848be975ebd1" (UID: "6d6e8c8f-6a59-443a-9874-848be975ebd1"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:30:38 crc kubenswrapper[4696]: I0321 09:30:38.822447 4696 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6d6e8c8f-6a59-443a-9874-848be975ebd1-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 21 09:30:40 crc kubenswrapper[4696]: I0321 09:30:40.546757 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d6e8c8f-6a59-443a-9874-848be975ebd1" path="/var/lib/kubelet/pods/6d6e8c8f-6a59-443a-9874-848be975ebd1/volumes" Mar 21 09:30:46 crc kubenswrapper[4696]: I0321 09:30:46.542118 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:30:46 crc kubenswrapper[4696]: E0321 09:30:46.542895 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:31:00 crc kubenswrapper[4696]: I0321 09:31:00.534994 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:31:00 crc kubenswrapper[4696]: E0321 09:31:00.535717 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:31:03 crc kubenswrapper[4696]: I0321 09:31:03.347895 4696 scope.go:117] "RemoveContainer" containerID="fbeec83987fec0cc2ddffebc516606011b51fcf88cc4f7355a71d4e239f53fe2" Mar 21 09:31:03 crc kubenswrapper[4696]: I0321 09:31:03.397801 4696 scope.go:117] "RemoveContainer" containerID="98aaa70452c44b76bf8ebe73361168b72c19f6b5beed5170a7ecbb55a3ee98d8" Mar 21 09:31:15 crc kubenswrapper[4696]: I0321 09:31:15.534844 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:31:15 crc kubenswrapper[4696]: E0321 09:31:15.535614 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:31:20 crc kubenswrapper[4696]: I0321 09:31:20.223388 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9g87s"] Mar 21 09:31:20 crc kubenswrapper[4696]: E0321 09:31:20.224472 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d6e8c8f-6a59-443a-9874-848be975ebd1" containerName="copy" Mar 21 09:31:20 crc kubenswrapper[4696]: I0321 09:31:20.224495 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d6e8c8f-6a59-443a-9874-848be975ebd1" containerName="copy" Mar 21 09:31:20 crc kubenswrapper[4696]: E0321 09:31:20.224555 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb1746c4-d2da-45f7-90a2-d28635623d57" containerName="oc" Mar 21 09:31:20 crc kubenswrapper[4696]: I0321 09:31:20.224565 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb1746c4-d2da-45f7-90a2-d28635623d57" containerName="oc" Mar 21 09:31:20 crc kubenswrapper[4696]: E0321 09:31:20.224582 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d6e8c8f-6a59-443a-9874-848be975ebd1" containerName="gather" Mar 21 09:31:20 crc kubenswrapper[4696]: I0321 09:31:20.224589 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d6e8c8f-6a59-443a-9874-848be975ebd1" containerName="gather" Mar 21 09:31:20 crc kubenswrapper[4696]: E0321 09:31:20.224599 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68000673-4e7e-4b9b-b45a-4c1172bc160b" containerName="collect-profiles" Mar 21 09:31:20 crc kubenswrapper[4696]: I0321 09:31:20.224606 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="68000673-4e7e-4b9b-b45a-4c1172bc160b" containerName="collect-profiles" Mar 21 09:31:20 crc kubenswrapper[4696]: I0321 09:31:20.224870 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="68000673-4e7e-4b9b-b45a-4c1172bc160b" containerName="collect-profiles" Mar 21 09:31:20 crc kubenswrapper[4696]: I0321 09:31:20.224887 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d6e8c8f-6a59-443a-9874-848be975ebd1" containerName="gather" Mar 21 09:31:20 crc kubenswrapper[4696]: I0321 09:31:20.224896 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d6e8c8f-6a59-443a-9874-848be975ebd1" containerName="copy" Mar 21 09:31:20 crc kubenswrapper[4696]: I0321 09:31:20.224908 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb1746c4-d2da-45f7-90a2-d28635623d57" containerName="oc" Mar 21 09:31:20 crc kubenswrapper[4696]: I0321 09:31:20.226869 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9g87s" Mar 21 09:31:20 crc kubenswrapper[4696]: I0321 09:31:20.237051 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9g87s"] Mar 21 09:31:20 crc kubenswrapper[4696]: I0321 09:31:20.356260 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a5555e7-fdad-4e0d-9713-80c9a23908cd-catalog-content\") pod \"community-operators-9g87s\" (UID: \"0a5555e7-fdad-4e0d-9713-80c9a23908cd\") " pod="openshift-marketplace/community-operators-9g87s" Mar 21 09:31:20 crc kubenswrapper[4696]: I0321 09:31:20.356623 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a5555e7-fdad-4e0d-9713-80c9a23908cd-utilities\") pod \"community-operators-9g87s\" (UID: \"0a5555e7-fdad-4e0d-9713-80c9a23908cd\") " pod="openshift-marketplace/community-operators-9g87s" Mar 21 09:31:20 crc kubenswrapper[4696]: I0321 09:31:20.356855 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv226\" (UniqueName: \"kubernetes.io/projected/0a5555e7-fdad-4e0d-9713-80c9a23908cd-kube-api-access-zv226\") pod \"community-operators-9g87s\" (UID: \"0a5555e7-fdad-4e0d-9713-80c9a23908cd\") " pod="openshift-marketplace/community-operators-9g87s" Mar 21 09:31:20 crc kubenswrapper[4696]: I0321 09:31:20.459080 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a5555e7-fdad-4e0d-9713-80c9a23908cd-utilities\") pod \"community-operators-9g87s\" (UID: \"0a5555e7-fdad-4e0d-9713-80c9a23908cd\") " pod="openshift-marketplace/community-operators-9g87s" Mar 21 09:31:20 crc kubenswrapper[4696]: I0321 09:31:20.459148 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv226\" (UniqueName: \"kubernetes.io/projected/0a5555e7-fdad-4e0d-9713-80c9a23908cd-kube-api-access-zv226\") pod \"community-operators-9g87s\" (UID: \"0a5555e7-fdad-4e0d-9713-80c9a23908cd\") " pod="openshift-marketplace/community-operators-9g87s" Mar 21 09:31:20 crc kubenswrapper[4696]: I0321 09:31:20.459253 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a5555e7-fdad-4e0d-9713-80c9a23908cd-catalog-content\") pod \"community-operators-9g87s\" (UID: \"0a5555e7-fdad-4e0d-9713-80c9a23908cd\") " pod="openshift-marketplace/community-operators-9g87s" Mar 21 09:31:20 crc kubenswrapper[4696]: I0321 09:31:20.459890 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a5555e7-fdad-4e0d-9713-80c9a23908cd-catalog-content\") pod \"community-operators-9g87s\" (UID: \"0a5555e7-fdad-4e0d-9713-80c9a23908cd\") " pod="openshift-marketplace/community-operators-9g87s" Mar 21 09:31:20 crc kubenswrapper[4696]: I0321 09:31:20.459971 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a5555e7-fdad-4e0d-9713-80c9a23908cd-utilities\") pod \"community-operators-9g87s\" (UID: \"0a5555e7-fdad-4e0d-9713-80c9a23908cd\") " pod="openshift-marketplace/community-operators-9g87s" Mar 21 09:31:20 crc kubenswrapper[4696]: I0321 09:31:20.477375 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv226\" (UniqueName: \"kubernetes.io/projected/0a5555e7-fdad-4e0d-9713-80c9a23908cd-kube-api-access-zv226\") pod \"community-operators-9g87s\" (UID: \"0a5555e7-fdad-4e0d-9713-80c9a23908cd\") " pod="openshift-marketplace/community-operators-9g87s" Mar 21 09:31:20 crc kubenswrapper[4696]: I0321 09:31:20.546676 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9g87s" Mar 21 09:31:21 crc kubenswrapper[4696]: I0321 09:31:21.317587 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9g87s"] Mar 21 09:31:21 crc kubenswrapper[4696]: I0321 09:31:21.905684 4696 generic.go:334] "Generic (PLEG): container finished" podID="0a5555e7-fdad-4e0d-9713-80c9a23908cd" containerID="a07a1acfc13ca369eeb90ce1594c8b752ac1903e7a24155c6248aa18282ac475" exitCode=0 Mar 21 09:31:21 crc kubenswrapper[4696]: I0321 09:31:21.905725 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9g87s" event={"ID":"0a5555e7-fdad-4e0d-9713-80c9a23908cd","Type":"ContainerDied","Data":"a07a1acfc13ca369eeb90ce1594c8b752ac1903e7a24155c6248aa18282ac475"} Mar 21 09:31:21 crc kubenswrapper[4696]: I0321 09:31:21.906003 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9g87s" event={"ID":"0a5555e7-fdad-4e0d-9713-80c9a23908cd","Type":"ContainerStarted","Data":"b138975cf522c4ac786cb7446f2701c9aff821eea85a33dfba89d821ffc1b856"} Mar 21 09:31:22 crc kubenswrapper[4696]: I0321 09:31:22.916252 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9g87s" event={"ID":"0a5555e7-fdad-4e0d-9713-80c9a23908cd","Type":"ContainerStarted","Data":"7229153b398462f82c92ffafe47b20aa3f1dc379e6896a14595763e86b1f8e01"} Mar 21 09:31:24 crc kubenswrapper[4696]: I0321 09:31:24.934497 4696 generic.go:334] "Generic (PLEG): container finished" podID="0a5555e7-fdad-4e0d-9713-80c9a23908cd" containerID="7229153b398462f82c92ffafe47b20aa3f1dc379e6896a14595763e86b1f8e01" exitCode=0 Mar 21 09:31:24 crc kubenswrapper[4696]: I0321 09:31:24.934584 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9g87s" event={"ID":"0a5555e7-fdad-4e0d-9713-80c9a23908cd","Type":"ContainerDied","Data":"7229153b398462f82c92ffafe47b20aa3f1dc379e6896a14595763e86b1f8e01"} Mar 21 09:31:25 crc kubenswrapper[4696]: I0321 09:31:25.947764 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9g87s" event={"ID":"0a5555e7-fdad-4e0d-9713-80c9a23908cd","Type":"ContainerStarted","Data":"4e5b0321fd530156724ea1a4d35c93e746b597a4380f4ebaa9f0404aedea0eb4"} Mar 21 09:31:25 crc kubenswrapper[4696]: I0321 09:31:25.973783 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9g87s" podStartSLOduration=2.535485254 podStartE2EDuration="5.973765832s" podCreationTimestamp="2026-03-21 09:31:20 +0000 UTC" firstStartedPulling="2026-03-21 09:31:21.907657941 +0000 UTC m=+3816.028538654" lastFinishedPulling="2026-03-21 09:31:25.345938519 +0000 UTC m=+3819.466819232" observedRunningTime="2026-03-21 09:31:25.970206918 +0000 UTC m=+3820.091087651" watchObservedRunningTime="2026-03-21 09:31:25.973765832 +0000 UTC m=+3820.094646545" Mar 21 09:31:27 crc kubenswrapper[4696]: I0321 09:31:27.535462 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:31:27 crc kubenswrapper[4696]: E0321 09:31:27.536372 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:31:30 crc kubenswrapper[4696]: I0321 09:31:30.546896 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9g87s" Mar 21 09:31:30 crc kubenswrapper[4696]: I0321 09:31:30.547220 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9g87s" Mar 21 09:31:30 crc kubenswrapper[4696]: I0321 09:31:30.599727 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9g87s" Mar 21 09:31:31 crc kubenswrapper[4696]: I0321 09:31:31.055057 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9g87s" Mar 21 09:31:31 crc kubenswrapper[4696]: I0321 09:31:31.844210 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9g87s"] Mar 21 09:31:33 crc kubenswrapper[4696]: I0321 09:31:33.017654 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9g87s" podUID="0a5555e7-fdad-4e0d-9713-80c9a23908cd" containerName="registry-server" containerID="cri-o://4e5b0321fd530156724ea1a4d35c93e746b597a4380f4ebaa9f0404aedea0eb4" gracePeriod=2 Mar 21 09:31:34 crc kubenswrapper[4696]: I0321 09:31:34.028443 4696 generic.go:334] "Generic (PLEG): container finished" podID="0a5555e7-fdad-4e0d-9713-80c9a23908cd" containerID="4e5b0321fd530156724ea1a4d35c93e746b597a4380f4ebaa9f0404aedea0eb4" exitCode=0 Mar 21 09:31:34 crc kubenswrapper[4696]: I0321 09:31:34.028514 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9g87s" event={"ID":"0a5555e7-fdad-4e0d-9713-80c9a23908cd","Type":"ContainerDied","Data":"4e5b0321fd530156724ea1a4d35c93e746b597a4380f4ebaa9f0404aedea0eb4"} Mar 21 09:31:34 crc kubenswrapper[4696]: I0321 09:31:34.318857 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9g87s" Mar 21 09:31:34 crc kubenswrapper[4696]: I0321 09:31:34.357914 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zv226\" (UniqueName: \"kubernetes.io/projected/0a5555e7-fdad-4e0d-9713-80c9a23908cd-kube-api-access-zv226\") pod \"0a5555e7-fdad-4e0d-9713-80c9a23908cd\" (UID: \"0a5555e7-fdad-4e0d-9713-80c9a23908cd\") " Mar 21 09:31:34 crc kubenswrapper[4696]: I0321 09:31:34.358160 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a5555e7-fdad-4e0d-9713-80c9a23908cd-catalog-content\") pod \"0a5555e7-fdad-4e0d-9713-80c9a23908cd\" (UID: \"0a5555e7-fdad-4e0d-9713-80c9a23908cd\") " Mar 21 09:31:34 crc kubenswrapper[4696]: I0321 09:31:34.358241 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a5555e7-fdad-4e0d-9713-80c9a23908cd-utilities\") pod \"0a5555e7-fdad-4e0d-9713-80c9a23908cd\" (UID: \"0a5555e7-fdad-4e0d-9713-80c9a23908cd\") " Mar 21 09:31:34 crc kubenswrapper[4696]: I0321 09:31:34.358932 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a5555e7-fdad-4e0d-9713-80c9a23908cd-utilities" (OuterVolumeSpecName: "utilities") pod "0a5555e7-fdad-4e0d-9713-80c9a23908cd" (UID: "0a5555e7-fdad-4e0d-9713-80c9a23908cd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:31:34 crc kubenswrapper[4696]: I0321 09:31:34.373430 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a5555e7-fdad-4e0d-9713-80c9a23908cd-kube-api-access-zv226" (OuterVolumeSpecName: "kube-api-access-zv226") pod "0a5555e7-fdad-4e0d-9713-80c9a23908cd" (UID: "0a5555e7-fdad-4e0d-9713-80c9a23908cd"). InnerVolumeSpecName "kube-api-access-zv226". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:31:34 crc kubenswrapper[4696]: I0321 09:31:34.441211 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a5555e7-fdad-4e0d-9713-80c9a23908cd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a5555e7-fdad-4e0d-9713-80c9a23908cd" (UID: "0a5555e7-fdad-4e0d-9713-80c9a23908cd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:31:34 crc kubenswrapper[4696]: I0321 09:31:34.463632 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zv226\" (UniqueName: \"kubernetes.io/projected/0a5555e7-fdad-4e0d-9713-80c9a23908cd-kube-api-access-zv226\") on node \"crc\" DevicePath \"\"" Mar 21 09:31:34 crc kubenswrapper[4696]: I0321 09:31:34.463661 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a5555e7-fdad-4e0d-9713-80c9a23908cd-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 09:31:34 crc kubenswrapper[4696]: I0321 09:31:34.463670 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a5555e7-fdad-4e0d-9713-80c9a23908cd-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 09:31:35 crc kubenswrapper[4696]: I0321 09:31:35.049112 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9g87s" event={"ID":"0a5555e7-fdad-4e0d-9713-80c9a23908cd","Type":"ContainerDied","Data":"b138975cf522c4ac786cb7446f2701c9aff821eea85a33dfba89d821ffc1b856"} Mar 21 09:31:35 crc kubenswrapper[4696]: I0321 09:31:35.049158 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9g87s" Mar 21 09:31:35 crc kubenswrapper[4696]: I0321 09:31:35.049177 4696 scope.go:117] "RemoveContainer" containerID="4e5b0321fd530156724ea1a4d35c93e746b597a4380f4ebaa9f0404aedea0eb4" Mar 21 09:31:35 crc kubenswrapper[4696]: I0321 09:31:35.096554 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9g87s"] Mar 21 09:31:35 crc kubenswrapper[4696]: I0321 09:31:35.098291 4696 scope.go:117] "RemoveContainer" containerID="7229153b398462f82c92ffafe47b20aa3f1dc379e6896a14595763e86b1f8e01" Mar 21 09:31:35 crc kubenswrapper[4696]: I0321 09:31:35.107692 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9g87s"] Mar 21 09:31:35 crc kubenswrapper[4696]: I0321 09:31:35.130439 4696 scope.go:117] "RemoveContainer" containerID="a07a1acfc13ca369eeb90ce1594c8b752ac1903e7a24155c6248aa18282ac475" Mar 21 09:31:36 crc kubenswrapper[4696]: I0321 09:31:36.546844 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a5555e7-fdad-4e0d-9713-80c9a23908cd" path="/var/lib/kubelet/pods/0a5555e7-fdad-4e0d-9713-80c9a23908cd/volumes" Mar 21 09:31:38 crc kubenswrapper[4696]: I0321 09:31:38.538106 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:31:38 crc kubenswrapper[4696]: E0321 09:31:38.538633 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:31:53 crc kubenswrapper[4696]: I0321 09:31:53.534557 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:31:53 crc kubenswrapper[4696]: E0321 09:31:53.535299 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:32:00 crc kubenswrapper[4696]: I0321 09:32:00.154153 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568092-brhlh"] Mar 21 09:32:00 crc kubenswrapper[4696]: E0321 09:32:00.155046 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a5555e7-fdad-4e0d-9713-80c9a23908cd" containerName="extract-content" Mar 21 09:32:00 crc kubenswrapper[4696]: I0321 09:32:00.155058 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a5555e7-fdad-4e0d-9713-80c9a23908cd" containerName="extract-content" Mar 21 09:32:00 crc kubenswrapper[4696]: E0321 09:32:00.155080 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a5555e7-fdad-4e0d-9713-80c9a23908cd" containerName="registry-server" Mar 21 09:32:00 crc kubenswrapper[4696]: I0321 09:32:00.155086 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a5555e7-fdad-4e0d-9713-80c9a23908cd" containerName="registry-server" Mar 21 09:32:00 crc kubenswrapper[4696]: E0321 09:32:00.155095 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a5555e7-fdad-4e0d-9713-80c9a23908cd" containerName="extract-utilities" Mar 21 09:32:00 crc kubenswrapper[4696]: I0321 09:32:00.155101 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a5555e7-fdad-4e0d-9713-80c9a23908cd" containerName="extract-utilities" Mar 21 09:32:00 crc kubenswrapper[4696]: I0321 09:32:00.155310 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a5555e7-fdad-4e0d-9713-80c9a23908cd" containerName="registry-server" Mar 21 09:32:00 crc kubenswrapper[4696]: I0321 09:32:00.156126 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568092-brhlh" Mar 21 09:32:00 crc kubenswrapper[4696]: I0321 09:32:00.161962 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:32:00 crc kubenswrapper[4696]: I0321 09:32:00.162143 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:32:00 crc kubenswrapper[4696]: I0321 09:32:00.162297 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:32:00 crc kubenswrapper[4696]: I0321 09:32:00.172270 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568092-brhlh"] Mar 21 09:32:00 crc kubenswrapper[4696]: I0321 09:32:00.210390 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8brsn\" (UniqueName: \"kubernetes.io/projected/9f4afa53-dca8-4e98-9067-841e281f49c9-kube-api-access-8brsn\") pod \"auto-csr-approver-29568092-brhlh\" (UID: \"9f4afa53-dca8-4e98-9067-841e281f49c9\") " pod="openshift-infra/auto-csr-approver-29568092-brhlh" Mar 21 09:32:00 crc kubenswrapper[4696]: I0321 09:32:00.312468 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8brsn\" (UniqueName: \"kubernetes.io/projected/9f4afa53-dca8-4e98-9067-841e281f49c9-kube-api-access-8brsn\") pod \"auto-csr-approver-29568092-brhlh\" (UID: \"9f4afa53-dca8-4e98-9067-841e281f49c9\") " pod="openshift-infra/auto-csr-approver-29568092-brhlh" Mar 21 09:32:00 crc kubenswrapper[4696]: I0321 09:32:00.331203 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8brsn\" (UniqueName: \"kubernetes.io/projected/9f4afa53-dca8-4e98-9067-841e281f49c9-kube-api-access-8brsn\") pod \"auto-csr-approver-29568092-brhlh\" (UID: \"9f4afa53-dca8-4e98-9067-841e281f49c9\") " pod="openshift-infra/auto-csr-approver-29568092-brhlh" Mar 21 09:32:00 crc kubenswrapper[4696]: I0321 09:32:00.486435 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568092-brhlh" Mar 21 09:32:01 crc kubenswrapper[4696]: I0321 09:32:01.203847 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568092-brhlh"] Mar 21 09:32:01 crc kubenswrapper[4696]: I0321 09:32:01.353271 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568092-brhlh" event={"ID":"9f4afa53-dca8-4e98-9067-841e281f49c9","Type":"ContainerStarted","Data":"a9e963d83f451198ee33e61fc0a3209fcbfb8ad6bc5d3dc81bd7609b303122fb"} Mar 21 09:32:02 crc kubenswrapper[4696]: I0321 09:32:02.363166 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568092-brhlh" event={"ID":"9f4afa53-dca8-4e98-9067-841e281f49c9","Type":"ContainerStarted","Data":"441f4831eadfbe70188580b2a40c0f5fe7dcdc96e70587660084294bcaa63505"} Mar 21 09:32:02 crc kubenswrapper[4696]: I0321 09:32:02.380701 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29568092-brhlh" podStartSLOduration=1.613838892 podStartE2EDuration="2.380686144s" podCreationTimestamp="2026-03-21 09:32:00 +0000 UTC" firstStartedPulling="2026-03-21 09:32:01.20850727 +0000 UTC m=+3855.329387983" lastFinishedPulling="2026-03-21 09:32:01.975354522 +0000 UTC m=+3856.096235235" observedRunningTime="2026-03-21 09:32:02.375512342 +0000 UTC m=+3856.496393055" watchObservedRunningTime="2026-03-21 09:32:02.380686144 +0000 UTC m=+3856.501566857" Mar 21 09:32:03 crc kubenswrapper[4696]: I0321 09:32:03.405161 4696 generic.go:334] "Generic (PLEG): container finished" podID="9f4afa53-dca8-4e98-9067-841e281f49c9" containerID="441f4831eadfbe70188580b2a40c0f5fe7dcdc96e70587660084294bcaa63505" exitCode=0 Mar 21 09:32:03 crc kubenswrapper[4696]: I0321 09:32:03.405467 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568092-brhlh" event={"ID":"9f4afa53-dca8-4e98-9067-841e281f49c9","Type":"ContainerDied","Data":"441f4831eadfbe70188580b2a40c0f5fe7dcdc96e70587660084294bcaa63505"} Mar 21 09:32:05 crc kubenswrapper[4696]: I0321 09:32:05.633665 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568092-brhlh" Mar 21 09:32:05 crc kubenswrapper[4696]: I0321 09:32:05.750431 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8brsn\" (UniqueName: \"kubernetes.io/projected/9f4afa53-dca8-4e98-9067-841e281f49c9-kube-api-access-8brsn\") pod \"9f4afa53-dca8-4e98-9067-841e281f49c9\" (UID: \"9f4afa53-dca8-4e98-9067-841e281f49c9\") " Mar 21 09:32:05 crc kubenswrapper[4696]: I0321 09:32:05.771972 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f4afa53-dca8-4e98-9067-841e281f49c9-kube-api-access-8brsn" (OuterVolumeSpecName: "kube-api-access-8brsn") pod "9f4afa53-dca8-4e98-9067-841e281f49c9" (UID: "9f4afa53-dca8-4e98-9067-841e281f49c9"). InnerVolumeSpecName "kube-api-access-8brsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:32:05 crc kubenswrapper[4696]: I0321 09:32:05.852283 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8brsn\" (UniqueName: \"kubernetes.io/projected/9f4afa53-dca8-4e98-9067-841e281f49c9-kube-api-access-8brsn\") on node \"crc\" DevicePath \"\"" Mar 21 09:32:06 crc kubenswrapper[4696]: I0321 09:32:06.436260 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568092-brhlh" event={"ID":"9f4afa53-dca8-4e98-9067-841e281f49c9","Type":"ContainerDied","Data":"a9e963d83f451198ee33e61fc0a3209fcbfb8ad6bc5d3dc81bd7609b303122fb"} Mar 21 09:32:06 crc kubenswrapper[4696]: I0321 09:32:06.436318 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9e963d83f451198ee33e61fc0a3209fcbfb8ad6bc5d3dc81bd7609b303122fb" Mar 21 09:32:06 crc kubenswrapper[4696]: I0321 09:32:06.436338 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568092-brhlh" Mar 21 09:32:06 crc kubenswrapper[4696]: I0321 09:32:06.704451 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568086-dx7lf"] Mar 21 09:32:06 crc kubenswrapper[4696]: I0321 09:32:06.715163 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568086-dx7lf"] Mar 21 09:32:08 crc kubenswrapper[4696]: I0321 09:32:08.535355 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:32:08 crc kubenswrapper[4696]: E0321 09:32:08.535911 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:32:08 crc kubenswrapper[4696]: I0321 09:32:08.545883 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57992ac0-359d-43ba-b9f3-d5e50e96cebd" path="/var/lib/kubelet/pods/57992ac0-359d-43ba-b9f3-d5e50e96cebd/volumes" Mar 21 09:32:20 crc kubenswrapper[4696]: I0321 09:32:20.535086 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:32:20 crc kubenswrapper[4696]: E0321 09:32:20.535898 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:32:35 crc kubenswrapper[4696]: I0321 09:32:35.535409 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:32:35 crc kubenswrapper[4696]: E0321 09:32:35.536184 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:32:46 crc kubenswrapper[4696]: I0321 09:32:46.541956 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:32:46 crc kubenswrapper[4696]: E0321 09:32:46.542855 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:32:57 crc kubenswrapper[4696]: I0321 09:32:57.534859 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:32:57 crc kubenswrapper[4696]: E0321 09:32:57.535609 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:33:03 crc kubenswrapper[4696]: I0321 09:33:03.540923 4696 scope.go:117] "RemoveContainer" containerID="c3af5de76bf59ba40cbb8323b2356ac81f1ac31ab6813d2cf714784f52f68387" Mar 21 09:33:08 crc kubenswrapper[4696]: I0321 09:33:08.534490 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:33:08 crc kubenswrapper[4696]: E0321 09:33:08.535134 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:33:19 crc kubenswrapper[4696]: I0321 09:33:19.534590 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:33:19 crc kubenswrapper[4696]: E0321 09:33:19.535360 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:33:33 crc kubenswrapper[4696]: I0321 09:33:33.534719 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:33:33 crc kubenswrapper[4696]: E0321 09:33:33.535686 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:33:47 crc kubenswrapper[4696]: I0321 09:33:47.535006 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:33:47 crc kubenswrapper[4696]: E0321 09:33:47.535787 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:33:50 crc kubenswrapper[4696]: I0321 09:33:50.873669 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gw6qx/must-gather-mn5rt"] Mar 21 09:33:50 crc kubenswrapper[4696]: E0321 09:33:50.874610 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f4afa53-dca8-4e98-9067-841e281f49c9" containerName="oc" Mar 21 09:33:50 crc kubenswrapper[4696]: I0321 09:33:50.874622 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f4afa53-dca8-4e98-9067-841e281f49c9" containerName="oc" Mar 21 09:33:50 crc kubenswrapper[4696]: I0321 09:33:50.874828 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f4afa53-dca8-4e98-9067-841e281f49c9" containerName="oc" Mar 21 09:33:50 crc kubenswrapper[4696]: I0321 09:33:50.875910 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gw6qx/must-gather-mn5rt" Mar 21 09:33:50 crc kubenswrapper[4696]: I0321 09:33:50.879835 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gw6qx"/"openshift-service-ca.crt" Mar 21 09:33:50 crc kubenswrapper[4696]: I0321 09:33:50.880052 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-gw6qx"/"default-dockercfg-wss9g" Mar 21 09:33:50 crc kubenswrapper[4696]: I0321 09:33:50.881058 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gw6qx"/"kube-root-ca.crt" Mar 21 09:33:50 crc kubenswrapper[4696]: I0321 09:33:50.889608 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gw6qx/must-gather-mn5rt"] Mar 21 09:33:50 crc kubenswrapper[4696]: I0321 09:33:50.962369 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f55d5321-1e8a-42bc-8c6e-17a2d032639b-must-gather-output\") pod \"must-gather-mn5rt\" (UID: \"f55d5321-1e8a-42bc-8c6e-17a2d032639b\") " pod="openshift-must-gather-gw6qx/must-gather-mn5rt" Mar 21 09:33:50 crc kubenswrapper[4696]: I0321 09:33:50.962592 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6wlz\" (UniqueName: \"kubernetes.io/projected/f55d5321-1e8a-42bc-8c6e-17a2d032639b-kube-api-access-j6wlz\") pod \"must-gather-mn5rt\" (UID: \"f55d5321-1e8a-42bc-8c6e-17a2d032639b\") " pod="openshift-must-gather-gw6qx/must-gather-mn5rt" Mar 21 09:33:51 crc kubenswrapper[4696]: I0321 09:33:51.064903 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6wlz\" (UniqueName: \"kubernetes.io/projected/f55d5321-1e8a-42bc-8c6e-17a2d032639b-kube-api-access-j6wlz\") pod \"must-gather-mn5rt\" (UID: \"f55d5321-1e8a-42bc-8c6e-17a2d032639b\") " pod="openshift-must-gather-gw6qx/must-gather-mn5rt" Mar 21 09:33:51 crc kubenswrapper[4696]: I0321 09:33:51.065372 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f55d5321-1e8a-42bc-8c6e-17a2d032639b-must-gather-output\") pod \"must-gather-mn5rt\" (UID: \"f55d5321-1e8a-42bc-8c6e-17a2d032639b\") " pod="openshift-must-gather-gw6qx/must-gather-mn5rt" Mar 21 09:33:51 crc kubenswrapper[4696]: I0321 09:33:51.065943 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f55d5321-1e8a-42bc-8c6e-17a2d032639b-must-gather-output\") pod \"must-gather-mn5rt\" (UID: \"f55d5321-1e8a-42bc-8c6e-17a2d032639b\") " pod="openshift-must-gather-gw6qx/must-gather-mn5rt" Mar 21 09:33:51 crc kubenswrapper[4696]: I0321 09:33:51.085086 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6wlz\" (UniqueName: \"kubernetes.io/projected/f55d5321-1e8a-42bc-8c6e-17a2d032639b-kube-api-access-j6wlz\") pod \"must-gather-mn5rt\" (UID: \"f55d5321-1e8a-42bc-8c6e-17a2d032639b\") " pod="openshift-must-gather-gw6qx/must-gather-mn5rt" Mar 21 09:33:51 crc kubenswrapper[4696]: I0321 09:33:51.208806 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gw6qx/must-gather-mn5rt" Mar 21 09:33:51 crc kubenswrapper[4696]: I0321 09:33:51.772236 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gw6qx/must-gather-mn5rt"] Mar 21 09:33:52 crc kubenswrapper[4696]: I0321 09:33:52.451731 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gw6qx/must-gather-mn5rt" event={"ID":"f55d5321-1e8a-42bc-8c6e-17a2d032639b","Type":"ContainerStarted","Data":"731bf94972a15a9640f391e81eb4eb80c2816d07d220cf2947b312112d371411"} Mar 21 09:33:52 crc kubenswrapper[4696]: I0321 09:33:52.452070 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gw6qx/must-gather-mn5rt" event={"ID":"f55d5321-1e8a-42bc-8c6e-17a2d032639b","Type":"ContainerStarted","Data":"33eae8dd16249faa0e0d07108fbe6627adf42580bb3a6c3cb04606153d5bd0b6"} Mar 21 09:33:52 crc kubenswrapper[4696]: I0321 09:33:52.452082 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gw6qx/must-gather-mn5rt" event={"ID":"f55d5321-1e8a-42bc-8c6e-17a2d032639b","Type":"ContainerStarted","Data":"e8ab384db6b50e2853edb3a464aefff77236947867c0e4ad81d51d5dc1a83955"} Mar 21 09:33:52 crc kubenswrapper[4696]: I0321 09:33:52.472145 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gw6qx/must-gather-mn5rt" podStartSLOduration=2.472127592 podStartE2EDuration="2.472127592s" podCreationTimestamp="2026-03-21 09:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 09:33:52.466402419 +0000 UTC m=+3966.587283132" watchObservedRunningTime="2026-03-21 09:33:52.472127592 +0000 UTC m=+3966.593008305" Mar 21 09:33:57 crc kubenswrapper[4696]: I0321 09:33:57.388630 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gw6qx/crc-debug-mnqfc"] Mar 21 09:33:57 crc kubenswrapper[4696]: I0321 09:33:57.390371 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gw6qx/crc-debug-mnqfc" Mar 21 09:33:57 crc kubenswrapper[4696]: I0321 09:33:57.423383 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbqft\" (UniqueName: \"kubernetes.io/projected/804ca533-d76d-4907-8876-ecc86d317f3d-kube-api-access-zbqft\") pod \"crc-debug-mnqfc\" (UID: \"804ca533-d76d-4907-8876-ecc86d317f3d\") " pod="openshift-must-gather-gw6qx/crc-debug-mnqfc" Mar 21 09:33:57 crc kubenswrapper[4696]: I0321 09:33:57.423625 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/804ca533-d76d-4907-8876-ecc86d317f3d-host\") pod \"crc-debug-mnqfc\" (UID: \"804ca533-d76d-4907-8876-ecc86d317f3d\") " pod="openshift-must-gather-gw6qx/crc-debug-mnqfc" Mar 21 09:33:57 crc kubenswrapper[4696]: I0321 09:33:57.525506 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbqft\" (UniqueName: \"kubernetes.io/projected/804ca533-d76d-4907-8876-ecc86d317f3d-kube-api-access-zbqft\") pod \"crc-debug-mnqfc\" (UID: \"804ca533-d76d-4907-8876-ecc86d317f3d\") " pod="openshift-must-gather-gw6qx/crc-debug-mnqfc" Mar 21 09:33:57 crc kubenswrapper[4696]: I0321 09:33:57.525637 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/804ca533-d76d-4907-8876-ecc86d317f3d-host\") pod \"crc-debug-mnqfc\" (UID: \"804ca533-d76d-4907-8876-ecc86d317f3d\") " pod="openshift-must-gather-gw6qx/crc-debug-mnqfc" Mar 21 09:33:57 crc kubenswrapper[4696]: I0321 09:33:57.525762 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/804ca533-d76d-4907-8876-ecc86d317f3d-host\") pod \"crc-debug-mnqfc\" (UID: \"804ca533-d76d-4907-8876-ecc86d317f3d\") " pod="openshift-must-gather-gw6qx/crc-debug-mnqfc" Mar 21 09:33:57 crc kubenswrapper[4696]: I0321 09:33:57.545584 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbqft\" (UniqueName: \"kubernetes.io/projected/804ca533-d76d-4907-8876-ecc86d317f3d-kube-api-access-zbqft\") pod \"crc-debug-mnqfc\" (UID: \"804ca533-d76d-4907-8876-ecc86d317f3d\") " pod="openshift-must-gather-gw6qx/crc-debug-mnqfc" Mar 21 09:33:57 crc kubenswrapper[4696]: I0321 09:33:57.708567 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gw6qx/crc-debug-mnqfc" Mar 21 09:33:58 crc kubenswrapper[4696]: I0321 09:33:58.509182 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gw6qx/crc-debug-mnqfc" event={"ID":"804ca533-d76d-4907-8876-ecc86d317f3d","Type":"ContainerStarted","Data":"2de4c3370996ba3ca8ec495d1be764876de2259f76adaca7b54ed497311aabbb"} Mar 21 09:33:58 crc kubenswrapper[4696]: I0321 09:33:58.509743 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gw6qx/crc-debug-mnqfc" event={"ID":"804ca533-d76d-4907-8876-ecc86d317f3d","Type":"ContainerStarted","Data":"dab615052ca1edc5fba69480353f8aa3b3467a87d46e363f2beb5195d8769fef"} Mar 21 09:33:58 crc kubenswrapper[4696]: I0321 09:33:58.529489 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gw6qx/crc-debug-mnqfc" podStartSLOduration=1.529471321 podStartE2EDuration="1.529471321s" podCreationTimestamp="2026-03-21 09:33:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 09:33:58.522224313 +0000 UTC m=+3972.643105036" watchObservedRunningTime="2026-03-21 09:33:58.529471321 +0000 UTC m=+3972.650352044" Mar 21 09:33:58 crc kubenswrapper[4696]: I0321 09:33:58.534412 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:33:58 crc kubenswrapper[4696]: E0321 09:33:58.534696 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:34:00 crc kubenswrapper[4696]: I0321 09:34:00.146553 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568094-92m6p"] Mar 21 09:34:00 crc kubenswrapper[4696]: I0321 09:34:00.148947 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568094-92m6p" Mar 21 09:34:00 crc kubenswrapper[4696]: I0321 09:34:00.153513 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:34:00 crc kubenswrapper[4696]: I0321 09:34:00.154261 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:34:00 crc kubenswrapper[4696]: I0321 09:34:00.158397 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:34:00 crc kubenswrapper[4696]: I0321 09:34:00.160614 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568094-92m6p"] Mar 21 09:34:00 crc kubenswrapper[4696]: I0321 09:34:00.181964 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b68tj\" (UniqueName: \"kubernetes.io/projected/0da3d36c-c518-48eb-8ea5-df44b5681206-kube-api-access-b68tj\") pod \"auto-csr-approver-29568094-92m6p\" (UID: \"0da3d36c-c518-48eb-8ea5-df44b5681206\") " pod="openshift-infra/auto-csr-approver-29568094-92m6p" Mar 21 09:34:00 crc kubenswrapper[4696]: I0321 09:34:00.284510 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b68tj\" (UniqueName: \"kubernetes.io/projected/0da3d36c-c518-48eb-8ea5-df44b5681206-kube-api-access-b68tj\") pod \"auto-csr-approver-29568094-92m6p\" (UID: \"0da3d36c-c518-48eb-8ea5-df44b5681206\") " pod="openshift-infra/auto-csr-approver-29568094-92m6p" Mar 21 09:34:00 crc kubenswrapper[4696]: I0321 09:34:00.308430 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b68tj\" (UniqueName: \"kubernetes.io/projected/0da3d36c-c518-48eb-8ea5-df44b5681206-kube-api-access-b68tj\") pod \"auto-csr-approver-29568094-92m6p\" (UID: \"0da3d36c-c518-48eb-8ea5-df44b5681206\") " pod="openshift-infra/auto-csr-approver-29568094-92m6p" Mar 21 09:34:00 crc kubenswrapper[4696]: I0321 09:34:00.472658 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568094-92m6p" Mar 21 09:34:01 crc kubenswrapper[4696]: I0321 09:34:01.229666 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568094-92m6p"] Mar 21 09:34:01 crc kubenswrapper[4696]: I0321 09:34:01.538829 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568094-92m6p" event={"ID":"0da3d36c-c518-48eb-8ea5-df44b5681206","Type":"ContainerStarted","Data":"a98f6451b5451e325cc5e28b1635b73b18c02275c7c2cec211499cab74b5b9b0"} Mar 21 09:34:01 crc kubenswrapper[4696]: I0321 09:34:01.927140 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-659lm"] Mar 21 09:34:01 crc kubenswrapper[4696]: I0321 09:34:01.930052 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-659lm" Mar 21 09:34:01 crc kubenswrapper[4696]: I0321 09:34:01.944341 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-659lm"] Mar 21 09:34:02 crc kubenswrapper[4696]: I0321 09:34:02.016587 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef37abbd-103e-44c3-846c-8505713468b7-catalog-content\") pod \"certified-operators-659lm\" (UID: \"ef37abbd-103e-44c3-846c-8505713468b7\") " pod="openshift-marketplace/certified-operators-659lm" Mar 21 09:34:02 crc kubenswrapper[4696]: I0321 09:34:02.016693 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef37abbd-103e-44c3-846c-8505713468b7-utilities\") pod \"certified-operators-659lm\" (UID: \"ef37abbd-103e-44c3-846c-8505713468b7\") " pod="openshift-marketplace/certified-operators-659lm" Mar 21 09:34:02 crc kubenswrapper[4696]: I0321 09:34:02.016744 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjwqf\" (UniqueName: \"kubernetes.io/projected/ef37abbd-103e-44c3-846c-8505713468b7-kube-api-access-mjwqf\") pod \"certified-operators-659lm\" (UID: \"ef37abbd-103e-44c3-846c-8505713468b7\") " pod="openshift-marketplace/certified-operators-659lm" Mar 21 09:34:02 crc kubenswrapper[4696]: I0321 09:34:02.119397 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef37abbd-103e-44c3-846c-8505713468b7-utilities\") pod \"certified-operators-659lm\" (UID: \"ef37abbd-103e-44c3-846c-8505713468b7\") " pod="openshift-marketplace/certified-operators-659lm" Mar 21 09:34:02 crc kubenswrapper[4696]: I0321 09:34:02.119542 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjwqf\" (UniqueName: \"kubernetes.io/projected/ef37abbd-103e-44c3-846c-8505713468b7-kube-api-access-mjwqf\") pod \"certified-operators-659lm\" (UID: \"ef37abbd-103e-44c3-846c-8505713468b7\") " pod="openshift-marketplace/certified-operators-659lm" Mar 21 09:34:02 crc kubenswrapper[4696]: I0321 09:34:02.119700 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef37abbd-103e-44c3-846c-8505713468b7-catalog-content\") pod \"certified-operators-659lm\" (UID: \"ef37abbd-103e-44c3-846c-8505713468b7\") " pod="openshift-marketplace/certified-operators-659lm" Mar 21 09:34:02 crc kubenswrapper[4696]: I0321 09:34:02.119850 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef37abbd-103e-44c3-846c-8505713468b7-utilities\") pod \"certified-operators-659lm\" (UID: \"ef37abbd-103e-44c3-846c-8505713468b7\") " pod="openshift-marketplace/certified-operators-659lm" Mar 21 09:34:02 crc kubenswrapper[4696]: I0321 09:34:02.120120 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef37abbd-103e-44c3-846c-8505713468b7-catalog-content\") pod \"certified-operators-659lm\" (UID: \"ef37abbd-103e-44c3-846c-8505713468b7\") " pod="openshift-marketplace/certified-operators-659lm" Mar 21 09:34:02 crc kubenswrapper[4696]: I0321 09:34:02.151749 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjwqf\" (UniqueName: \"kubernetes.io/projected/ef37abbd-103e-44c3-846c-8505713468b7-kube-api-access-mjwqf\") pod \"certified-operators-659lm\" (UID: \"ef37abbd-103e-44c3-846c-8505713468b7\") " pod="openshift-marketplace/certified-operators-659lm" Mar 21 09:34:02 crc kubenswrapper[4696]: I0321 09:34:02.263450 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-659lm" Mar 21 09:34:02 crc kubenswrapper[4696]: I0321 09:34:02.567111 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568094-92m6p" event={"ID":"0da3d36c-c518-48eb-8ea5-df44b5681206","Type":"ContainerStarted","Data":"46461fd2615b0d286720ddf99d31e4e1206699d9f8b5dc98fadfed16e5f9aec4"} Mar 21 09:34:02 crc kubenswrapper[4696]: I0321 09:34:02.598949 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29568094-92m6p" podStartSLOduration=1.698757197 podStartE2EDuration="2.598928691s" podCreationTimestamp="2026-03-21 09:34:00 +0000 UTC" firstStartedPulling="2026-03-21 09:34:01.22004976 +0000 UTC m=+3975.340930473" lastFinishedPulling="2026-03-21 09:34:02.120221254 +0000 UTC m=+3976.241101967" observedRunningTime="2026-03-21 09:34:02.591800546 +0000 UTC m=+3976.712681259" watchObservedRunningTime="2026-03-21 09:34:02.598928691 +0000 UTC m=+3976.719809404" Mar 21 09:34:02 crc kubenswrapper[4696]: I0321 09:34:02.839323 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-659lm"] Mar 21 09:34:03 crc kubenswrapper[4696]: I0321 09:34:03.577984 4696 generic.go:334] "Generic (PLEG): container finished" podID="ef37abbd-103e-44c3-846c-8505713468b7" containerID="decabe93fb45b8172b01a09f70f0a2570fb6942bee2021734c1401f4ec7c0dcc" exitCode=0 Mar 21 09:34:03 crc kubenswrapper[4696]: I0321 09:34:03.578316 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-659lm" event={"ID":"ef37abbd-103e-44c3-846c-8505713468b7","Type":"ContainerDied","Data":"decabe93fb45b8172b01a09f70f0a2570fb6942bee2021734c1401f4ec7c0dcc"} Mar 21 09:34:03 crc kubenswrapper[4696]: I0321 09:34:03.578343 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-659lm" event={"ID":"ef37abbd-103e-44c3-846c-8505713468b7","Type":"ContainerStarted","Data":"1bc8c758f8580c3cee9ed1e02d04b40987289a2fe00bf1446a4d1e3ddf12020e"} Mar 21 09:34:03 crc kubenswrapper[4696]: I0321 09:34:03.582628 4696 generic.go:334] "Generic (PLEG): container finished" podID="0da3d36c-c518-48eb-8ea5-df44b5681206" containerID="46461fd2615b0d286720ddf99d31e4e1206699d9f8b5dc98fadfed16e5f9aec4" exitCode=0 Mar 21 09:34:03 crc kubenswrapper[4696]: I0321 09:34:03.582663 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568094-92m6p" event={"ID":"0da3d36c-c518-48eb-8ea5-df44b5681206","Type":"ContainerDied","Data":"46461fd2615b0d286720ddf99d31e4e1206699d9f8b5dc98fadfed16e5f9aec4"} Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.118093 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-czz7c"] Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.120915 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-czz7c" Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.136085 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-czz7c"] Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.197283 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ba0054-0957-49e9-8a32-876ba4f4b987-utilities\") pod \"redhat-marketplace-czz7c\" (UID: \"17ba0054-0957-49e9-8a32-876ba4f4b987\") " pod="openshift-marketplace/redhat-marketplace-czz7c" Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.197545 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ba0054-0957-49e9-8a32-876ba4f4b987-catalog-content\") pod \"redhat-marketplace-czz7c\" (UID: \"17ba0054-0957-49e9-8a32-876ba4f4b987\") " pod="openshift-marketplace/redhat-marketplace-czz7c" Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.197698 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzbb5\" (UniqueName: \"kubernetes.io/projected/17ba0054-0957-49e9-8a32-876ba4f4b987-kube-api-access-vzbb5\") pod \"redhat-marketplace-czz7c\" (UID: \"17ba0054-0957-49e9-8a32-876ba4f4b987\") " pod="openshift-marketplace/redhat-marketplace-czz7c" Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.300279 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ba0054-0957-49e9-8a32-876ba4f4b987-utilities\") pod \"redhat-marketplace-czz7c\" (UID: \"17ba0054-0957-49e9-8a32-876ba4f4b987\") " pod="openshift-marketplace/redhat-marketplace-czz7c" Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.300640 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ba0054-0957-49e9-8a32-876ba4f4b987-catalog-content\") pod \"redhat-marketplace-czz7c\" (UID: \"17ba0054-0957-49e9-8a32-876ba4f4b987\") " pod="openshift-marketplace/redhat-marketplace-czz7c" Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.300905 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ba0054-0957-49e9-8a32-876ba4f4b987-utilities\") pod \"redhat-marketplace-czz7c\" (UID: \"17ba0054-0957-49e9-8a32-876ba4f4b987\") " pod="openshift-marketplace/redhat-marketplace-czz7c" Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.300924 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzbb5\" (UniqueName: \"kubernetes.io/projected/17ba0054-0957-49e9-8a32-876ba4f4b987-kube-api-access-vzbb5\") pod \"redhat-marketplace-czz7c\" (UID: \"17ba0054-0957-49e9-8a32-876ba4f4b987\") " pod="openshift-marketplace/redhat-marketplace-czz7c" Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.301130 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ba0054-0957-49e9-8a32-876ba4f4b987-catalog-content\") pod \"redhat-marketplace-czz7c\" (UID: \"17ba0054-0957-49e9-8a32-876ba4f4b987\") " pod="openshift-marketplace/redhat-marketplace-czz7c" Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.321833 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzbb5\" (UniqueName: \"kubernetes.io/projected/17ba0054-0957-49e9-8a32-876ba4f4b987-kube-api-access-vzbb5\") pod \"redhat-marketplace-czz7c\" (UID: \"17ba0054-0957-49e9-8a32-876ba4f4b987\") " pod="openshift-marketplace/redhat-marketplace-czz7c" Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.466881 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-czz7c" Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.561165 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568094-92m6p" Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.608207 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b68tj\" (UniqueName: \"kubernetes.io/projected/0da3d36c-c518-48eb-8ea5-df44b5681206-kube-api-access-b68tj\") pod \"0da3d36c-c518-48eb-8ea5-df44b5681206\" (UID: \"0da3d36c-c518-48eb-8ea5-df44b5681206\") " Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.638004 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0da3d36c-c518-48eb-8ea5-df44b5681206-kube-api-access-b68tj" (OuterVolumeSpecName: "kube-api-access-b68tj") pod "0da3d36c-c518-48eb-8ea5-df44b5681206" (UID: "0da3d36c-c518-48eb-8ea5-df44b5681206"). InnerVolumeSpecName "kube-api-access-b68tj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.647107 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568094-92m6p" event={"ID":"0da3d36c-c518-48eb-8ea5-df44b5681206","Type":"ContainerDied","Data":"a98f6451b5451e325cc5e28b1635b73b18c02275c7c2cec211499cab74b5b9b0"} Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.647149 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a98f6451b5451e325cc5e28b1635b73b18c02275c7c2cec211499cab74b5b9b0" Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.647221 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568094-92m6p" Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.659986 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-659lm" event={"ID":"ef37abbd-103e-44c3-846c-8505713468b7","Type":"ContainerStarted","Data":"8583f1b01e3aabd39e3f827322dc9fa23315dc1449b27409cfef0b0000b2af43"} Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.704010 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568088-w2ckn"] Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.711217 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b68tj\" (UniqueName: \"kubernetes.io/projected/0da3d36c-c518-48eb-8ea5-df44b5681206-kube-api-access-b68tj\") on node \"crc\" DevicePath \"\"" Mar 21 09:34:05 crc kubenswrapper[4696]: I0321 09:34:05.721227 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568088-w2ckn"] Mar 21 09:34:06 crc kubenswrapper[4696]: I0321 09:34:06.231356 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-czz7c"] Mar 21 09:34:06 crc kubenswrapper[4696]: I0321 09:34:06.547895 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3a7dacd-a824-4d86-bf9b-57c1cfa220df" path="/var/lib/kubelet/pods/b3a7dacd-a824-4d86-bf9b-57c1cfa220df/volumes" Mar 21 09:34:06 crc kubenswrapper[4696]: W0321 09:34:06.775917 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17ba0054_0957_49e9_8a32_876ba4f4b987.slice/crio-3c9ad3f105cb558144e27e9037a0b8742c0c07f1abcaf4a4c4c591b0f12875fa WatchSource:0}: Error finding container 3c9ad3f105cb558144e27e9037a0b8742c0c07f1abcaf4a4c4c591b0f12875fa: Status 404 returned error can't find the container with id 3c9ad3f105cb558144e27e9037a0b8742c0c07f1abcaf4a4c4c591b0f12875fa Mar 21 09:34:07 crc kubenswrapper[4696]: I0321 09:34:07.693867 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czz7c" event={"ID":"17ba0054-0957-49e9-8a32-876ba4f4b987","Type":"ContainerStarted","Data":"3c9ad3f105cb558144e27e9037a0b8742c0c07f1abcaf4a4c4c591b0f12875fa"} Mar 21 09:34:07 crc kubenswrapper[4696]: I0321 09:34:07.696739 4696 generic.go:334] "Generic (PLEG): container finished" podID="ef37abbd-103e-44c3-846c-8505713468b7" containerID="8583f1b01e3aabd39e3f827322dc9fa23315dc1449b27409cfef0b0000b2af43" exitCode=0 Mar 21 09:34:07 crc kubenswrapper[4696]: I0321 09:34:07.696784 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-659lm" event={"ID":"ef37abbd-103e-44c3-846c-8505713468b7","Type":"ContainerDied","Data":"8583f1b01e3aabd39e3f827322dc9fa23315dc1449b27409cfef0b0000b2af43"} Mar 21 09:34:08 crc kubenswrapper[4696]: I0321 09:34:08.721145 4696 generic.go:334] "Generic (PLEG): container finished" podID="17ba0054-0957-49e9-8a32-876ba4f4b987" containerID="9883203e663db62ec9ef4b42bfc76f4d97cbedb172227e93b8f0b89ffb0a72cd" exitCode=0 Mar 21 09:34:08 crc kubenswrapper[4696]: I0321 09:34:08.722316 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czz7c" event={"ID":"17ba0054-0957-49e9-8a32-876ba4f4b987","Type":"ContainerDied","Data":"9883203e663db62ec9ef4b42bfc76f4d97cbedb172227e93b8f0b89ffb0a72cd"} Mar 21 09:34:08 crc kubenswrapper[4696]: I0321 09:34:08.727731 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-659lm" event={"ID":"ef37abbd-103e-44c3-846c-8505713468b7","Type":"ContainerStarted","Data":"7d002de19a8a8cf85aeb95b15de750983dd2f6652f23d4d57a310812771966d8"} Mar 21 09:34:09 crc kubenswrapper[4696]: I0321 09:34:09.741506 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czz7c" event={"ID":"17ba0054-0957-49e9-8a32-876ba4f4b987","Type":"ContainerStarted","Data":"520e45f892d998ba05a89db45b717c566ab1e6c039b8a1a0902960b68a391256"} Mar 21 09:34:09 crc kubenswrapper[4696]: I0321 09:34:09.771521 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-659lm" podStartSLOduration=4.002565423 podStartE2EDuration="8.771503291s" podCreationTimestamp="2026-03-21 09:34:01 +0000 UTC" firstStartedPulling="2026-03-21 09:34:03.580252558 +0000 UTC m=+3977.701133271" lastFinishedPulling="2026-03-21 09:34:08.349190426 +0000 UTC m=+3982.470071139" observedRunningTime="2026-03-21 09:34:08.775226005 +0000 UTC m=+3982.896106748" watchObservedRunningTime="2026-03-21 09:34:09.771503291 +0000 UTC m=+3983.892384004" Mar 21 09:34:11 crc kubenswrapper[4696]: I0321 09:34:11.761419 4696 generic.go:334] "Generic (PLEG): container finished" podID="17ba0054-0957-49e9-8a32-876ba4f4b987" containerID="520e45f892d998ba05a89db45b717c566ab1e6c039b8a1a0902960b68a391256" exitCode=0 Mar 21 09:34:11 crc kubenswrapper[4696]: I0321 09:34:11.761504 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czz7c" event={"ID":"17ba0054-0957-49e9-8a32-876ba4f4b987","Type":"ContainerDied","Data":"520e45f892d998ba05a89db45b717c566ab1e6c039b8a1a0902960b68a391256"} Mar 21 09:34:12 crc kubenswrapper[4696]: I0321 09:34:12.264468 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-659lm" Mar 21 09:34:12 crc kubenswrapper[4696]: I0321 09:34:12.264772 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-659lm" Mar 21 09:34:12 crc kubenswrapper[4696]: I0321 09:34:12.320932 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-659lm" Mar 21 09:34:12 crc kubenswrapper[4696]: I0321 09:34:12.535308 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:34:12 crc kubenswrapper[4696]: E0321 09:34:12.535885 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:34:12 crc kubenswrapper[4696]: I0321 09:34:12.773186 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czz7c" event={"ID":"17ba0054-0957-49e9-8a32-876ba4f4b987","Type":"ContainerStarted","Data":"8338ad4703d22b6bd59548fb354a04e025e7cd58efb4cef501fd91a8c0196c4b"} Mar 21 09:34:12 crc kubenswrapper[4696]: I0321 09:34:12.795388 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-czz7c" podStartSLOduration=4.358597228 podStartE2EDuration="7.795365552s" podCreationTimestamp="2026-03-21 09:34:05 +0000 UTC" firstStartedPulling="2026-03-21 09:34:08.724828882 +0000 UTC m=+3982.845709595" lastFinishedPulling="2026-03-21 09:34:12.161597206 +0000 UTC m=+3986.282477919" observedRunningTime="2026-03-21 09:34:12.787845087 +0000 UTC m=+3986.908725820" watchObservedRunningTime="2026-03-21 09:34:12.795365552 +0000 UTC m=+3986.916246265" Mar 21 09:34:15 crc kubenswrapper[4696]: I0321 09:34:15.467894 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-czz7c" Mar 21 09:34:15 crc kubenswrapper[4696]: I0321 09:34:15.468248 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-czz7c" Mar 21 09:34:16 crc kubenswrapper[4696]: I0321 09:34:16.517760 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-czz7c" podUID="17ba0054-0957-49e9-8a32-876ba4f4b987" containerName="registry-server" probeResult="failure" output=< Mar 21 09:34:16 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Mar 21 09:34:16 crc kubenswrapper[4696]: > Mar 21 09:34:22 crc kubenswrapper[4696]: I0321 09:34:22.315361 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-659lm" Mar 21 09:34:22 crc kubenswrapper[4696]: I0321 09:34:22.711508 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-659lm"] Mar 21 09:34:22 crc kubenswrapper[4696]: I0321 09:34:22.874626 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-659lm" podUID="ef37abbd-103e-44c3-846c-8505713468b7" containerName="registry-server" containerID="cri-o://7d002de19a8a8cf85aeb95b15de750983dd2f6652f23d4d57a310812771966d8" gracePeriod=2 Mar 21 09:34:23 crc kubenswrapper[4696]: I0321 09:34:23.763452 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-659lm" Mar 21 09:34:23 crc kubenswrapper[4696]: I0321 09:34:23.840870 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef37abbd-103e-44c3-846c-8505713468b7-utilities\") pod \"ef37abbd-103e-44c3-846c-8505713468b7\" (UID: \"ef37abbd-103e-44c3-846c-8505713468b7\") " Mar 21 09:34:23 crc kubenswrapper[4696]: I0321 09:34:23.840935 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef37abbd-103e-44c3-846c-8505713468b7-catalog-content\") pod \"ef37abbd-103e-44c3-846c-8505713468b7\" (UID: \"ef37abbd-103e-44c3-846c-8505713468b7\") " Mar 21 09:34:23 crc kubenswrapper[4696]: I0321 09:34:23.841091 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjwqf\" (UniqueName: \"kubernetes.io/projected/ef37abbd-103e-44c3-846c-8505713468b7-kube-api-access-mjwqf\") pod \"ef37abbd-103e-44c3-846c-8505713468b7\" (UID: \"ef37abbd-103e-44c3-846c-8505713468b7\") " Mar 21 09:34:23 crc kubenswrapper[4696]: I0321 09:34:23.841533 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef37abbd-103e-44c3-846c-8505713468b7-utilities" (OuterVolumeSpecName: "utilities") pod "ef37abbd-103e-44c3-846c-8505713468b7" (UID: "ef37abbd-103e-44c3-846c-8505713468b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:34:23 crc kubenswrapper[4696]: I0321 09:34:23.846578 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef37abbd-103e-44c3-846c-8505713468b7-kube-api-access-mjwqf" (OuterVolumeSpecName: "kube-api-access-mjwqf") pod "ef37abbd-103e-44c3-846c-8505713468b7" (UID: "ef37abbd-103e-44c3-846c-8505713468b7"). InnerVolumeSpecName "kube-api-access-mjwqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:34:23 crc kubenswrapper[4696]: I0321 09:34:23.889751 4696 generic.go:334] "Generic (PLEG): container finished" podID="ef37abbd-103e-44c3-846c-8505713468b7" containerID="7d002de19a8a8cf85aeb95b15de750983dd2f6652f23d4d57a310812771966d8" exitCode=0 Mar 21 09:34:23 crc kubenswrapper[4696]: I0321 09:34:23.889840 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-659lm" event={"ID":"ef37abbd-103e-44c3-846c-8505713468b7","Type":"ContainerDied","Data":"7d002de19a8a8cf85aeb95b15de750983dd2f6652f23d4d57a310812771966d8"} Mar 21 09:34:23 crc kubenswrapper[4696]: I0321 09:34:23.889880 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-659lm" event={"ID":"ef37abbd-103e-44c3-846c-8505713468b7","Type":"ContainerDied","Data":"1bc8c758f8580c3cee9ed1e02d04b40987289a2fe00bf1446a4d1e3ddf12020e"} Mar 21 09:34:23 crc kubenswrapper[4696]: I0321 09:34:23.889899 4696 scope.go:117] "RemoveContainer" containerID="7d002de19a8a8cf85aeb95b15de750983dd2f6652f23d4d57a310812771966d8" Mar 21 09:34:23 crc kubenswrapper[4696]: I0321 09:34:23.890075 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-659lm" Mar 21 09:34:23 crc kubenswrapper[4696]: I0321 09:34:23.901223 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef37abbd-103e-44c3-846c-8505713468b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ef37abbd-103e-44c3-846c-8505713468b7" (UID: "ef37abbd-103e-44c3-846c-8505713468b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:34:23 crc kubenswrapper[4696]: I0321 09:34:23.920339 4696 scope.go:117] "RemoveContainer" containerID="8583f1b01e3aabd39e3f827322dc9fa23315dc1449b27409cfef0b0000b2af43" Mar 21 09:34:23 crc kubenswrapper[4696]: I0321 09:34:23.944262 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef37abbd-103e-44c3-846c-8505713468b7-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 09:34:23 crc kubenswrapper[4696]: I0321 09:34:23.944304 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef37abbd-103e-44c3-846c-8505713468b7-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 09:34:23 crc kubenswrapper[4696]: I0321 09:34:23.944320 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjwqf\" (UniqueName: \"kubernetes.io/projected/ef37abbd-103e-44c3-846c-8505713468b7-kube-api-access-mjwqf\") on node \"crc\" DevicePath \"\"" Mar 21 09:34:23 crc kubenswrapper[4696]: I0321 09:34:23.951789 4696 scope.go:117] "RemoveContainer" containerID="decabe93fb45b8172b01a09f70f0a2570fb6942bee2021734c1401f4ec7c0dcc" Mar 21 09:34:24 crc kubenswrapper[4696]: I0321 09:34:24.222596 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-659lm"] Mar 21 09:34:24 crc kubenswrapper[4696]: I0321 09:34:24.232006 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-659lm"] Mar 21 09:34:24 crc kubenswrapper[4696]: I0321 09:34:24.386717 4696 scope.go:117] "RemoveContainer" containerID="7d002de19a8a8cf85aeb95b15de750983dd2f6652f23d4d57a310812771966d8" Mar 21 09:34:24 crc kubenswrapper[4696]: E0321 09:34:24.387301 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d002de19a8a8cf85aeb95b15de750983dd2f6652f23d4d57a310812771966d8\": container with ID starting with 7d002de19a8a8cf85aeb95b15de750983dd2f6652f23d4d57a310812771966d8 not found: ID does not exist" containerID="7d002de19a8a8cf85aeb95b15de750983dd2f6652f23d4d57a310812771966d8" Mar 21 09:34:24 crc kubenswrapper[4696]: I0321 09:34:24.387423 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d002de19a8a8cf85aeb95b15de750983dd2f6652f23d4d57a310812771966d8"} err="failed to get container status \"7d002de19a8a8cf85aeb95b15de750983dd2f6652f23d4d57a310812771966d8\": rpc error: code = NotFound desc = could not find container \"7d002de19a8a8cf85aeb95b15de750983dd2f6652f23d4d57a310812771966d8\": container with ID starting with 7d002de19a8a8cf85aeb95b15de750983dd2f6652f23d4d57a310812771966d8 not found: ID does not exist" Mar 21 09:34:24 crc kubenswrapper[4696]: I0321 09:34:24.387511 4696 scope.go:117] "RemoveContainer" containerID="8583f1b01e3aabd39e3f827322dc9fa23315dc1449b27409cfef0b0000b2af43" Mar 21 09:34:24 crc kubenswrapper[4696]: E0321 09:34:24.388090 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8583f1b01e3aabd39e3f827322dc9fa23315dc1449b27409cfef0b0000b2af43\": container with ID starting with 8583f1b01e3aabd39e3f827322dc9fa23315dc1449b27409cfef0b0000b2af43 not found: ID does not exist" containerID="8583f1b01e3aabd39e3f827322dc9fa23315dc1449b27409cfef0b0000b2af43" Mar 21 09:34:24 crc kubenswrapper[4696]: I0321 09:34:24.388131 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8583f1b01e3aabd39e3f827322dc9fa23315dc1449b27409cfef0b0000b2af43"} err="failed to get container status \"8583f1b01e3aabd39e3f827322dc9fa23315dc1449b27409cfef0b0000b2af43\": rpc error: code = NotFound desc = could not find container \"8583f1b01e3aabd39e3f827322dc9fa23315dc1449b27409cfef0b0000b2af43\": container with ID starting with 8583f1b01e3aabd39e3f827322dc9fa23315dc1449b27409cfef0b0000b2af43 not found: ID does not exist" Mar 21 09:34:24 crc kubenswrapper[4696]: I0321 09:34:24.388160 4696 scope.go:117] "RemoveContainer" containerID="decabe93fb45b8172b01a09f70f0a2570fb6942bee2021734c1401f4ec7c0dcc" Mar 21 09:34:24 crc kubenswrapper[4696]: E0321 09:34:24.388415 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"decabe93fb45b8172b01a09f70f0a2570fb6942bee2021734c1401f4ec7c0dcc\": container with ID starting with decabe93fb45b8172b01a09f70f0a2570fb6942bee2021734c1401f4ec7c0dcc not found: ID does not exist" containerID="decabe93fb45b8172b01a09f70f0a2570fb6942bee2021734c1401f4ec7c0dcc" Mar 21 09:34:24 crc kubenswrapper[4696]: I0321 09:34:24.388446 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"decabe93fb45b8172b01a09f70f0a2570fb6942bee2021734c1401f4ec7c0dcc"} err="failed to get container status \"decabe93fb45b8172b01a09f70f0a2570fb6942bee2021734c1401f4ec7c0dcc\": rpc error: code = NotFound desc = could not find container \"decabe93fb45b8172b01a09f70f0a2570fb6942bee2021734c1401f4ec7c0dcc\": container with ID starting with decabe93fb45b8172b01a09f70f0a2570fb6942bee2021734c1401f4ec7c0dcc not found: ID does not exist" Mar 21 09:34:24 crc kubenswrapper[4696]: I0321 09:34:24.535141 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:34:24 crc kubenswrapper[4696]: E0321 09:34:24.535548 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:34:24 crc kubenswrapper[4696]: I0321 09:34:24.547520 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef37abbd-103e-44c3-846c-8505713468b7" path="/var/lib/kubelet/pods/ef37abbd-103e-44c3-846c-8505713468b7/volumes" Mar 21 09:34:25 crc kubenswrapper[4696]: I0321 09:34:25.528975 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-czz7c" Mar 21 09:34:25 crc kubenswrapper[4696]: I0321 09:34:25.580461 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-czz7c" Mar 21 09:34:27 crc kubenswrapper[4696]: I0321 09:34:27.111563 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-czz7c"] Mar 21 09:34:27 crc kubenswrapper[4696]: I0321 09:34:27.112185 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-czz7c" podUID="17ba0054-0957-49e9-8a32-876ba4f4b987" containerName="registry-server" containerID="cri-o://8338ad4703d22b6bd59548fb354a04e025e7cd58efb4cef501fd91a8c0196c4b" gracePeriod=2 Mar 21 09:34:27 crc kubenswrapper[4696]: I0321 09:34:27.932786 4696 generic.go:334] "Generic (PLEG): container finished" podID="17ba0054-0957-49e9-8a32-876ba4f4b987" containerID="8338ad4703d22b6bd59548fb354a04e025e7cd58efb4cef501fd91a8c0196c4b" exitCode=0 Mar 21 09:34:27 crc kubenswrapper[4696]: I0321 09:34:27.932940 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czz7c" event={"ID":"17ba0054-0957-49e9-8a32-876ba4f4b987","Type":"ContainerDied","Data":"8338ad4703d22b6bd59548fb354a04e025e7cd58efb4cef501fd91a8c0196c4b"} Mar 21 09:34:27 crc kubenswrapper[4696]: I0321 09:34:27.933120 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czz7c" event={"ID":"17ba0054-0957-49e9-8a32-876ba4f4b987","Type":"ContainerDied","Data":"3c9ad3f105cb558144e27e9037a0b8742c0c07f1abcaf4a4c4c591b0f12875fa"} Mar 21 09:34:27 crc kubenswrapper[4696]: I0321 09:34:27.933137 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c9ad3f105cb558144e27e9037a0b8742c0c07f1abcaf4a4c4c591b0f12875fa" Mar 21 09:34:27 crc kubenswrapper[4696]: I0321 09:34:27.951747 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-czz7c" Mar 21 09:34:28 crc kubenswrapper[4696]: I0321 09:34:28.035669 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ba0054-0957-49e9-8a32-876ba4f4b987-utilities\") pod \"17ba0054-0957-49e9-8a32-876ba4f4b987\" (UID: \"17ba0054-0957-49e9-8a32-876ba4f4b987\") " Mar 21 09:34:28 crc kubenswrapper[4696]: I0321 09:34:28.035741 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzbb5\" (UniqueName: \"kubernetes.io/projected/17ba0054-0957-49e9-8a32-876ba4f4b987-kube-api-access-vzbb5\") pod \"17ba0054-0957-49e9-8a32-876ba4f4b987\" (UID: \"17ba0054-0957-49e9-8a32-876ba4f4b987\") " Mar 21 09:34:28 crc kubenswrapper[4696]: I0321 09:34:28.035901 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ba0054-0957-49e9-8a32-876ba4f4b987-catalog-content\") pod \"17ba0054-0957-49e9-8a32-876ba4f4b987\" (UID: \"17ba0054-0957-49e9-8a32-876ba4f4b987\") " Mar 21 09:34:28 crc kubenswrapper[4696]: I0321 09:34:28.039151 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17ba0054-0957-49e9-8a32-876ba4f4b987-utilities" (OuterVolumeSpecName: "utilities") pod "17ba0054-0957-49e9-8a32-876ba4f4b987" (UID: "17ba0054-0957-49e9-8a32-876ba4f4b987"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:34:28 crc kubenswrapper[4696]: I0321 09:34:28.044005 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17ba0054-0957-49e9-8a32-876ba4f4b987-kube-api-access-vzbb5" (OuterVolumeSpecName: "kube-api-access-vzbb5") pod "17ba0054-0957-49e9-8a32-876ba4f4b987" (UID: "17ba0054-0957-49e9-8a32-876ba4f4b987"). InnerVolumeSpecName "kube-api-access-vzbb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:34:28 crc kubenswrapper[4696]: I0321 09:34:28.084374 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17ba0054-0957-49e9-8a32-876ba4f4b987-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17ba0054-0957-49e9-8a32-876ba4f4b987" (UID: "17ba0054-0957-49e9-8a32-876ba4f4b987"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:34:28 crc kubenswrapper[4696]: I0321 09:34:28.139540 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ba0054-0957-49e9-8a32-876ba4f4b987-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 09:34:28 crc kubenswrapper[4696]: I0321 09:34:28.139591 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzbb5\" (UniqueName: \"kubernetes.io/projected/17ba0054-0957-49e9-8a32-876ba4f4b987-kube-api-access-vzbb5\") on node \"crc\" DevicePath \"\"" Mar 21 09:34:28 crc kubenswrapper[4696]: I0321 09:34:28.139609 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ba0054-0957-49e9-8a32-876ba4f4b987-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 09:34:28 crc kubenswrapper[4696]: I0321 09:34:28.942714 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-czz7c" Mar 21 09:34:28 crc kubenswrapper[4696]: I0321 09:34:28.970849 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-czz7c"] Mar 21 09:34:28 crc kubenswrapper[4696]: I0321 09:34:28.984813 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-czz7c"] Mar 21 09:34:30 crc kubenswrapper[4696]: I0321 09:34:30.545657 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17ba0054-0957-49e9-8a32-876ba4f4b987" path="/var/lib/kubelet/pods/17ba0054-0957-49e9-8a32-876ba4f4b987/volumes" Mar 21 09:34:37 crc kubenswrapper[4696]: I0321 09:34:37.535398 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:34:38 crc kubenswrapper[4696]: I0321 09:34:38.029278 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerStarted","Data":"1967ae7e219ab85d3941d84ab56ee52381e3cd6376788e863ea1aa5183ce99b5"} Mar 21 09:34:48 crc kubenswrapper[4696]: I0321 09:34:48.126754 4696 generic.go:334] "Generic (PLEG): container finished" podID="804ca533-d76d-4907-8876-ecc86d317f3d" containerID="2de4c3370996ba3ca8ec495d1be764876de2259f76adaca7b54ed497311aabbb" exitCode=0 Mar 21 09:34:48 crc kubenswrapper[4696]: I0321 09:34:48.126802 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gw6qx/crc-debug-mnqfc" event={"ID":"804ca533-d76d-4907-8876-ecc86d317f3d","Type":"ContainerDied","Data":"2de4c3370996ba3ca8ec495d1be764876de2259f76adaca7b54ed497311aabbb"} Mar 21 09:34:49 crc kubenswrapper[4696]: I0321 09:34:49.264919 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gw6qx/crc-debug-mnqfc" Mar 21 09:34:49 crc kubenswrapper[4696]: I0321 09:34:49.268459 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/804ca533-d76d-4907-8876-ecc86d317f3d-host\") pod \"804ca533-d76d-4907-8876-ecc86d317f3d\" (UID: \"804ca533-d76d-4907-8876-ecc86d317f3d\") " Mar 21 09:34:49 crc kubenswrapper[4696]: I0321 09:34:49.268526 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/804ca533-d76d-4907-8876-ecc86d317f3d-host" (OuterVolumeSpecName: "host") pod "804ca533-d76d-4907-8876-ecc86d317f3d" (UID: "804ca533-d76d-4907-8876-ecc86d317f3d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 09:34:49 crc kubenswrapper[4696]: I0321 09:34:49.268621 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbqft\" (UniqueName: \"kubernetes.io/projected/804ca533-d76d-4907-8876-ecc86d317f3d-kube-api-access-zbqft\") pod \"804ca533-d76d-4907-8876-ecc86d317f3d\" (UID: \"804ca533-d76d-4907-8876-ecc86d317f3d\") " Mar 21 09:34:49 crc kubenswrapper[4696]: I0321 09:34:49.269016 4696 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/804ca533-d76d-4907-8876-ecc86d317f3d-host\") on node \"crc\" DevicePath \"\"" Mar 21 09:34:49 crc kubenswrapper[4696]: I0321 09:34:49.281441 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/804ca533-d76d-4907-8876-ecc86d317f3d-kube-api-access-zbqft" (OuterVolumeSpecName: "kube-api-access-zbqft") pod "804ca533-d76d-4907-8876-ecc86d317f3d" (UID: "804ca533-d76d-4907-8876-ecc86d317f3d"). InnerVolumeSpecName "kube-api-access-zbqft". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:34:49 crc kubenswrapper[4696]: I0321 09:34:49.309458 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gw6qx/crc-debug-mnqfc"] Mar 21 09:34:49 crc kubenswrapper[4696]: I0321 09:34:49.325808 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gw6qx/crc-debug-mnqfc"] Mar 21 09:34:49 crc kubenswrapper[4696]: I0321 09:34:49.370341 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbqft\" (UniqueName: \"kubernetes.io/projected/804ca533-d76d-4907-8876-ecc86d317f3d-kube-api-access-zbqft\") on node \"crc\" DevicePath \"\"" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.144285 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dab615052ca1edc5fba69480353f8aa3b3467a87d46e363f2beb5195d8769fef" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.144318 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gw6qx/crc-debug-mnqfc" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.501788 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gw6qx/crc-debug-7tcmt"] Mar 21 09:34:50 crc kubenswrapper[4696]: E0321 09:34:50.502219 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ba0054-0957-49e9-8a32-876ba4f4b987" containerName="registry-server" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.502232 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ba0054-0957-49e9-8a32-876ba4f4b987" containerName="registry-server" Mar 21 09:34:50 crc kubenswrapper[4696]: E0321 09:34:50.502244 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ba0054-0957-49e9-8a32-876ba4f4b987" containerName="extract-utilities" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.502250 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ba0054-0957-49e9-8a32-876ba4f4b987" containerName="extract-utilities" Mar 21 09:34:50 crc kubenswrapper[4696]: E0321 09:34:50.502272 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0da3d36c-c518-48eb-8ea5-df44b5681206" containerName="oc" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.502278 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="0da3d36c-c518-48eb-8ea5-df44b5681206" containerName="oc" Mar 21 09:34:50 crc kubenswrapper[4696]: E0321 09:34:50.502290 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="804ca533-d76d-4907-8876-ecc86d317f3d" containerName="container-00" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.502295 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="804ca533-d76d-4907-8876-ecc86d317f3d" containerName="container-00" Mar 21 09:34:50 crc kubenswrapper[4696]: E0321 09:34:50.502324 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef37abbd-103e-44c3-846c-8505713468b7" containerName="extract-utilities" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.502330 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef37abbd-103e-44c3-846c-8505713468b7" containerName="extract-utilities" Mar 21 09:34:50 crc kubenswrapper[4696]: E0321 09:34:50.502339 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ba0054-0957-49e9-8a32-876ba4f4b987" containerName="extract-content" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.502346 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ba0054-0957-49e9-8a32-876ba4f4b987" containerName="extract-content" Mar 21 09:34:50 crc kubenswrapper[4696]: E0321 09:34:50.502360 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef37abbd-103e-44c3-846c-8505713468b7" containerName="extract-content" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.502366 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef37abbd-103e-44c3-846c-8505713468b7" containerName="extract-content" Mar 21 09:34:50 crc kubenswrapper[4696]: E0321 09:34:50.502377 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef37abbd-103e-44c3-846c-8505713468b7" containerName="registry-server" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.502384 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef37abbd-103e-44c3-846c-8505713468b7" containerName="registry-server" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.502559 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="804ca533-d76d-4907-8876-ecc86d317f3d" containerName="container-00" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.502575 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef37abbd-103e-44c3-846c-8505713468b7" containerName="registry-server" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.502587 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="17ba0054-0957-49e9-8a32-876ba4f4b987" containerName="registry-server" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.502780 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="0da3d36c-c518-48eb-8ea5-df44b5681206" containerName="oc" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.503486 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gw6qx/crc-debug-7tcmt" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.544890 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="804ca533-d76d-4907-8876-ecc86d317f3d" path="/var/lib/kubelet/pods/804ca533-d76d-4907-8876-ecc86d317f3d/volumes" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.596899 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2kfc\" (UniqueName: \"kubernetes.io/projected/2f397869-67a8-4c88-b764-e75eef59ab19-kube-api-access-j2kfc\") pod \"crc-debug-7tcmt\" (UID: \"2f397869-67a8-4c88-b764-e75eef59ab19\") " pod="openshift-must-gather-gw6qx/crc-debug-7tcmt" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.597116 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2f397869-67a8-4c88-b764-e75eef59ab19-host\") pod \"crc-debug-7tcmt\" (UID: \"2f397869-67a8-4c88-b764-e75eef59ab19\") " pod="openshift-must-gather-gw6qx/crc-debug-7tcmt" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.699286 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2kfc\" (UniqueName: \"kubernetes.io/projected/2f397869-67a8-4c88-b764-e75eef59ab19-kube-api-access-j2kfc\") pod \"crc-debug-7tcmt\" (UID: \"2f397869-67a8-4c88-b764-e75eef59ab19\") " pod="openshift-must-gather-gw6qx/crc-debug-7tcmt" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.699434 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2f397869-67a8-4c88-b764-e75eef59ab19-host\") pod \"crc-debug-7tcmt\" (UID: \"2f397869-67a8-4c88-b764-e75eef59ab19\") " pod="openshift-must-gather-gw6qx/crc-debug-7tcmt" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.699572 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2f397869-67a8-4c88-b764-e75eef59ab19-host\") pod \"crc-debug-7tcmt\" (UID: \"2f397869-67a8-4c88-b764-e75eef59ab19\") " pod="openshift-must-gather-gw6qx/crc-debug-7tcmt" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.716474 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2kfc\" (UniqueName: \"kubernetes.io/projected/2f397869-67a8-4c88-b764-e75eef59ab19-kube-api-access-j2kfc\") pod \"crc-debug-7tcmt\" (UID: \"2f397869-67a8-4c88-b764-e75eef59ab19\") " pod="openshift-must-gather-gw6qx/crc-debug-7tcmt" Mar 21 09:34:50 crc kubenswrapper[4696]: I0321 09:34:50.824219 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gw6qx/crc-debug-7tcmt" Mar 21 09:34:50 crc kubenswrapper[4696]: W0321 09:34:50.859893 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f397869_67a8_4c88_b764_e75eef59ab19.slice/crio-49aa7c5993381a71d48a3ca9df926bd90218a3d7920504e48e09b0e3bf23a1c1 WatchSource:0}: Error finding container 49aa7c5993381a71d48a3ca9df926bd90218a3d7920504e48e09b0e3bf23a1c1: Status 404 returned error can't find the container with id 49aa7c5993381a71d48a3ca9df926bd90218a3d7920504e48e09b0e3bf23a1c1 Mar 21 09:34:51 crc kubenswrapper[4696]: I0321 09:34:51.154226 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gw6qx/crc-debug-7tcmt" event={"ID":"2f397869-67a8-4c88-b764-e75eef59ab19","Type":"ContainerStarted","Data":"50f64b41930601c15e20c2cfcb0ef014df74cb5f1d6808c3a37a76a9639e45db"} Mar 21 09:34:51 crc kubenswrapper[4696]: I0321 09:34:51.154576 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gw6qx/crc-debug-7tcmt" event={"ID":"2f397869-67a8-4c88-b764-e75eef59ab19","Type":"ContainerStarted","Data":"49aa7c5993381a71d48a3ca9df926bd90218a3d7920504e48e09b0e3bf23a1c1"} Mar 21 09:34:51 crc kubenswrapper[4696]: I0321 09:34:51.166325 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gw6qx/crc-debug-7tcmt" podStartSLOduration=1.166307018 podStartE2EDuration="1.166307018s" podCreationTimestamp="2026-03-21 09:34:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 09:34:51.165947058 +0000 UTC m=+4025.286827771" watchObservedRunningTime="2026-03-21 09:34:51.166307018 +0000 UTC m=+4025.287187731" Mar 21 09:34:52 crc kubenswrapper[4696]: I0321 09:34:52.171659 4696 generic.go:334] "Generic (PLEG): container finished" podID="2f397869-67a8-4c88-b764-e75eef59ab19" containerID="50f64b41930601c15e20c2cfcb0ef014df74cb5f1d6808c3a37a76a9639e45db" exitCode=0 Mar 21 09:34:52 crc kubenswrapper[4696]: I0321 09:34:52.171985 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gw6qx/crc-debug-7tcmt" event={"ID":"2f397869-67a8-4c88-b764-e75eef59ab19","Type":"ContainerDied","Data":"50f64b41930601c15e20c2cfcb0ef014df74cb5f1d6808c3a37a76a9639e45db"} Mar 21 09:34:53 crc kubenswrapper[4696]: I0321 09:34:53.287120 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gw6qx/crc-debug-7tcmt" Mar 21 09:34:53 crc kubenswrapper[4696]: I0321 09:34:53.332294 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gw6qx/crc-debug-7tcmt"] Mar 21 09:34:53 crc kubenswrapper[4696]: I0321 09:34:53.344886 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gw6qx/crc-debug-7tcmt"] Mar 21 09:34:53 crc kubenswrapper[4696]: I0321 09:34:53.356238 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2kfc\" (UniqueName: \"kubernetes.io/projected/2f397869-67a8-4c88-b764-e75eef59ab19-kube-api-access-j2kfc\") pod \"2f397869-67a8-4c88-b764-e75eef59ab19\" (UID: \"2f397869-67a8-4c88-b764-e75eef59ab19\") " Mar 21 09:34:53 crc kubenswrapper[4696]: I0321 09:34:53.356523 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2f397869-67a8-4c88-b764-e75eef59ab19-host\") pod \"2f397869-67a8-4c88-b764-e75eef59ab19\" (UID: \"2f397869-67a8-4c88-b764-e75eef59ab19\") " Mar 21 09:34:53 crc kubenswrapper[4696]: I0321 09:34:53.356647 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f397869-67a8-4c88-b764-e75eef59ab19-host" (OuterVolumeSpecName: "host") pod "2f397869-67a8-4c88-b764-e75eef59ab19" (UID: "2f397869-67a8-4c88-b764-e75eef59ab19"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 09:34:53 crc kubenswrapper[4696]: I0321 09:34:53.357288 4696 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2f397869-67a8-4c88-b764-e75eef59ab19-host\") on node \"crc\" DevicePath \"\"" Mar 21 09:34:53 crc kubenswrapper[4696]: I0321 09:34:53.362987 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f397869-67a8-4c88-b764-e75eef59ab19-kube-api-access-j2kfc" (OuterVolumeSpecName: "kube-api-access-j2kfc") pod "2f397869-67a8-4c88-b764-e75eef59ab19" (UID: "2f397869-67a8-4c88-b764-e75eef59ab19"). InnerVolumeSpecName "kube-api-access-j2kfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:34:53 crc kubenswrapper[4696]: I0321 09:34:53.459456 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2kfc\" (UniqueName: \"kubernetes.io/projected/2f397869-67a8-4c88-b764-e75eef59ab19-kube-api-access-j2kfc\") on node \"crc\" DevicePath \"\"" Mar 21 09:34:54 crc kubenswrapper[4696]: I0321 09:34:54.189744 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49aa7c5993381a71d48a3ca9df926bd90218a3d7920504e48e09b0e3bf23a1c1" Mar 21 09:34:54 crc kubenswrapper[4696]: I0321 09:34:54.189813 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gw6qx/crc-debug-7tcmt" Mar 21 09:34:54 crc kubenswrapper[4696]: I0321 09:34:54.478586 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gw6qx/crc-debug-dpr62"] Mar 21 09:34:54 crc kubenswrapper[4696]: E0321 09:34:54.478995 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f397869-67a8-4c88-b764-e75eef59ab19" containerName="container-00" Mar 21 09:34:54 crc kubenswrapper[4696]: I0321 09:34:54.479009 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f397869-67a8-4c88-b764-e75eef59ab19" containerName="container-00" Mar 21 09:34:54 crc kubenswrapper[4696]: I0321 09:34:54.479238 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f397869-67a8-4c88-b764-e75eef59ab19" containerName="container-00" Mar 21 09:34:54 crc kubenswrapper[4696]: I0321 09:34:54.480006 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gw6qx/crc-debug-dpr62" Mar 21 09:34:54 crc kubenswrapper[4696]: I0321 09:34:54.545882 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f397869-67a8-4c88-b764-e75eef59ab19" path="/var/lib/kubelet/pods/2f397869-67a8-4c88-b764-e75eef59ab19/volumes" Mar 21 09:34:54 crc kubenswrapper[4696]: I0321 09:34:54.634241 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc8wl\" (UniqueName: \"kubernetes.io/projected/4eceb525-7807-4676-8da7-0f7b74ce478c-kube-api-access-jc8wl\") pod \"crc-debug-dpr62\" (UID: \"4eceb525-7807-4676-8da7-0f7b74ce478c\") " pod="openshift-must-gather-gw6qx/crc-debug-dpr62" Mar 21 09:34:54 crc kubenswrapper[4696]: I0321 09:34:54.634661 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4eceb525-7807-4676-8da7-0f7b74ce478c-host\") pod \"crc-debug-dpr62\" (UID: \"4eceb525-7807-4676-8da7-0f7b74ce478c\") " pod="openshift-must-gather-gw6qx/crc-debug-dpr62" Mar 21 09:34:54 crc kubenswrapper[4696]: I0321 09:34:54.736075 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc8wl\" (UniqueName: \"kubernetes.io/projected/4eceb525-7807-4676-8da7-0f7b74ce478c-kube-api-access-jc8wl\") pod \"crc-debug-dpr62\" (UID: \"4eceb525-7807-4676-8da7-0f7b74ce478c\") " pod="openshift-must-gather-gw6qx/crc-debug-dpr62" Mar 21 09:34:54 crc kubenswrapper[4696]: I0321 09:34:54.736261 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4eceb525-7807-4676-8da7-0f7b74ce478c-host\") pod \"crc-debug-dpr62\" (UID: \"4eceb525-7807-4676-8da7-0f7b74ce478c\") " pod="openshift-must-gather-gw6qx/crc-debug-dpr62" Mar 21 09:34:54 crc kubenswrapper[4696]: I0321 09:34:54.736411 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4eceb525-7807-4676-8da7-0f7b74ce478c-host\") pod \"crc-debug-dpr62\" (UID: \"4eceb525-7807-4676-8da7-0f7b74ce478c\") " pod="openshift-must-gather-gw6qx/crc-debug-dpr62" Mar 21 09:34:54 crc kubenswrapper[4696]: I0321 09:34:54.753158 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc8wl\" (UniqueName: \"kubernetes.io/projected/4eceb525-7807-4676-8da7-0f7b74ce478c-kube-api-access-jc8wl\") pod \"crc-debug-dpr62\" (UID: \"4eceb525-7807-4676-8da7-0f7b74ce478c\") " pod="openshift-must-gather-gw6qx/crc-debug-dpr62" Mar 21 09:34:54 crc kubenswrapper[4696]: I0321 09:34:54.796339 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gw6qx/crc-debug-dpr62" Mar 21 09:34:55 crc kubenswrapper[4696]: I0321 09:34:55.202655 4696 generic.go:334] "Generic (PLEG): container finished" podID="4eceb525-7807-4676-8da7-0f7b74ce478c" containerID="90fba04cea61cb932678f6d634063fa8e3dc002687b77e4b7df2f36152dfb33c" exitCode=0 Mar 21 09:34:55 crc kubenswrapper[4696]: I0321 09:34:55.202718 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gw6qx/crc-debug-dpr62" event={"ID":"4eceb525-7807-4676-8da7-0f7b74ce478c","Type":"ContainerDied","Data":"90fba04cea61cb932678f6d634063fa8e3dc002687b77e4b7df2f36152dfb33c"} Mar 21 09:34:55 crc kubenswrapper[4696]: I0321 09:34:55.202978 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gw6qx/crc-debug-dpr62" event={"ID":"4eceb525-7807-4676-8da7-0f7b74ce478c","Type":"ContainerStarted","Data":"977610fc6ea1dc64e77f2c4083f819562f3697efc58ab04cf8827044a9c5d04b"} Mar 21 09:34:55 crc kubenswrapper[4696]: I0321 09:34:55.250417 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gw6qx/crc-debug-dpr62"] Mar 21 09:34:55 crc kubenswrapper[4696]: I0321 09:34:55.260370 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gw6qx/crc-debug-dpr62"] Mar 21 09:34:56 crc kubenswrapper[4696]: I0321 09:34:56.806991 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gw6qx/crc-debug-dpr62" Mar 21 09:34:56 crc kubenswrapper[4696]: I0321 09:34:56.876769 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4eceb525-7807-4676-8da7-0f7b74ce478c-host\") pod \"4eceb525-7807-4676-8da7-0f7b74ce478c\" (UID: \"4eceb525-7807-4676-8da7-0f7b74ce478c\") " Mar 21 09:34:56 crc kubenswrapper[4696]: I0321 09:34:56.876885 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jc8wl\" (UniqueName: \"kubernetes.io/projected/4eceb525-7807-4676-8da7-0f7b74ce478c-kube-api-access-jc8wl\") pod \"4eceb525-7807-4676-8da7-0f7b74ce478c\" (UID: \"4eceb525-7807-4676-8da7-0f7b74ce478c\") " Mar 21 09:34:56 crc kubenswrapper[4696]: I0321 09:34:56.877294 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4eceb525-7807-4676-8da7-0f7b74ce478c-host" (OuterVolumeSpecName: "host") pod "4eceb525-7807-4676-8da7-0f7b74ce478c" (UID: "4eceb525-7807-4676-8da7-0f7b74ce478c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 21 09:34:56 crc kubenswrapper[4696]: I0321 09:34:56.877765 4696 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4eceb525-7807-4676-8da7-0f7b74ce478c-host\") on node \"crc\" DevicePath \"\"" Mar 21 09:34:56 crc kubenswrapper[4696]: I0321 09:34:56.882838 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eceb525-7807-4676-8da7-0f7b74ce478c-kube-api-access-jc8wl" (OuterVolumeSpecName: "kube-api-access-jc8wl") pod "4eceb525-7807-4676-8da7-0f7b74ce478c" (UID: "4eceb525-7807-4676-8da7-0f7b74ce478c"). InnerVolumeSpecName "kube-api-access-jc8wl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:34:56 crc kubenswrapper[4696]: I0321 09:34:56.979483 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jc8wl\" (UniqueName: \"kubernetes.io/projected/4eceb525-7807-4676-8da7-0f7b74ce478c-kube-api-access-jc8wl\") on node \"crc\" DevicePath \"\"" Mar 21 09:34:57 crc kubenswrapper[4696]: I0321 09:34:57.232564 4696 scope.go:117] "RemoveContainer" containerID="90fba04cea61cb932678f6d634063fa8e3dc002687b77e4b7df2f36152dfb33c" Mar 21 09:34:57 crc kubenswrapper[4696]: I0321 09:34:57.232585 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gw6qx/crc-debug-dpr62" Mar 21 09:34:58 crc kubenswrapper[4696]: I0321 09:34:58.544330 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4eceb525-7807-4676-8da7-0f7b74ce478c" path="/var/lib/kubelet/pods/4eceb525-7807-4676-8da7-0f7b74ce478c/volumes" Mar 21 09:35:03 crc kubenswrapper[4696]: I0321 09:35:03.639240 4696 scope.go:117] "RemoveContainer" containerID="19e31703b8381b0f31d344e849892b111ffc7efa6e5589d7a48a416d6db45179" Mar 21 09:35:40 crc kubenswrapper[4696]: I0321 09:35:40.125649 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_b1808d81-bf6f-4201-972a-507d6b41f33f/init-config-reloader/0.log" Mar 21 09:35:40 crc kubenswrapper[4696]: I0321 09:35:40.426967 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_b1808d81-bf6f-4201-972a-507d6b41f33f/alertmanager/0.log" Mar 21 09:35:40 crc kubenswrapper[4696]: I0321 09:35:40.436013 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_b1808d81-bf6f-4201-972a-507d6b41f33f/config-reloader/0.log" Mar 21 09:35:40 crc kubenswrapper[4696]: I0321 09:35:40.472293 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_b1808d81-bf6f-4201-972a-507d6b41f33f/init-config-reloader/0.log" Mar 21 09:35:40 crc kubenswrapper[4696]: I0321 09:35:40.604325 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7dd88fbdb6-nnxhg_bd0e5187-2172-48fc-90e8-4ce8f3f1c143/barbican-api-log/0.log" Mar 21 09:35:40 crc kubenswrapper[4696]: I0321 09:35:40.635096 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7dd88fbdb6-nnxhg_bd0e5187-2172-48fc-90e8-4ce8f3f1c143/barbican-api/0.log" Mar 21 09:35:40 crc kubenswrapper[4696]: I0321 09:35:40.756529 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7c894bbd86-9phxx_adeff71a-2b97-46e9-a945-d69cf06e34d9/barbican-keystone-listener/0.log" Mar 21 09:35:40 crc kubenswrapper[4696]: I0321 09:35:40.922082 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7c894bbd86-9phxx_adeff71a-2b97-46e9-a945-d69cf06e34d9/barbican-keystone-listener-log/0.log" Mar 21 09:35:40 crc kubenswrapper[4696]: I0321 09:35:40.946915 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-756466fd6c-6vdmf_639973de-e4aa-46d4-817a-f9c410238b45/barbican-worker/0.log" Mar 21 09:35:40 crc kubenswrapper[4696]: I0321 09:35:40.985299 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-756466fd6c-6vdmf_639973de-e4aa-46d4-817a-f9c410238b45/barbican-worker-log/0.log" Mar 21 09:35:41 crc kubenswrapper[4696]: I0321 09:35:41.230501 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-k8rlq_875486e6-3369-4601-bc53-00c2aaf85707/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:35:41 crc kubenswrapper[4696]: I0321 09:35:41.263927 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_72bd0225-bcbe-44ec-be2a-4f38093ea9c9/ceilometer-central-agent/0.log" Mar 21 09:35:41 crc kubenswrapper[4696]: I0321 09:35:41.356635 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_72bd0225-bcbe-44ec-be2a-4f38093ea9c9/ceilometer-notification-agent/0.log" Mar 21 09:35:41 crc kubenswrapper[4696]: I0321 09:35:41.414919 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_72bd0225-bcbe-44ec-be2a-4f38093ea9c9/proxy-httpd/0.log" Mar 21 09:35:41 crc kubenswrapper[4696]: I0321 09:35:41.443471 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_72bd0225-bcbe-44ec-be2a-4f38093ea9c9/sg-core/0.log" Mar 21 09:35:41 crc kubenswrapper[4696]: I0321 09:35:41.602290 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8d228678-e5cf-4da4-9a96-7ed39576c142/cinder-api/0.log" Mar 21 09:35:41 crc kubenswrapper[4696]: I0321 09:35:41.628791 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8d228678-e5cf-4da4-9a96-7ed39576c142/cinder-api-log/0.log" Mar 21 09:35:41 crc kubenswrapper[4696]: I0321 09:35:41.830056 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4fce4b85-2f39-4702-b221-fd5e870e6a2d/cinder-scheduler/0.log" Mar 21 09:35:41 crc kubenswrapper[4696]: I0321 09:35:41.840976 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4fce4b85-2f39-4702-b221-fd5e870e6a2d/probe/0.log" Mar 21 09:35:41 crc kubenswrapper[4696]: I0321 09:35:41.955598 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_16a5bae2-5738-4f3c-a6ca-1e08aa8930a0/cloudkitty-api/0.log" Mar 21 09:35:42 crc kubenswrapper[4696]: I0321 09:35:42.071249 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_16a5bae2-5738-4f3c-a6ca-1e08aa8930a0/cloudkitty-api-log/0.log" Mar 21 09:35:42 crc kubenswrapper[4696]: I0321 09:35:42.107325 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-compactor-0_14c7fb5d-c3dd-4751-b2a2-b4cc618cd7a8/loki-compactor/0.log" Mar 21 09:35:42 crc kubenswrapper[4696]: I0321 09:35:42.289435 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-distributor-5d547bbd4d-rxzh7_a3c11f23-8050-4ea7-ab4c-4b9866ce0ff7/loki-distributor/0.log" Mar 21 09:35:42 crc kubenswrapper[4696]: I0321 09:35:42.337708 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-6b884dc4b5-lczl8_d2b0b7e1-1fa0-44ce-bab0-0964460e645b/gateway/0.log" Mar 21 09:35:42 crc kubenswrapper[4696]: I0321 09:35:42.474738 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-6b884dc4b5-qgq6w_70a97941-5dea-48a7-8dbf-b475b4af0b6e/gateway/0.log" Mar 21 09:35:42 crc kubenswrapper[4696]: I0321 09:35:42.581241 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-index-gateway-0_88865f3d-9bb9-4844-bb96-dc25c11615ec/loki-index-gateway/0.log" Mar 21 09:35:42 crc kubenswrapper[4696]: I0321 09:35:42.789141 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-ingester-0_d6518eba-34ec-4a38-b279-cd8bb11e7e24/loki-ingester/0.log" Mar 21 09:35:43 crc kubenswrapper[4696]: I0321 09:35:43.046752 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-query-frontend-6f54889599-6mn7k_bbf61ae0-6918-44af-be69-ac8a220fcd6b/loki-query-frontend/0.log" Mar 21 09:35:43 crc kubenswrapper[4696]: I0321 09:35:43.312254 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-querier-668f98fdd7-tnh77_edf07829-775b-4672-89fe-81c3342300ae/loki-querier/0.log" Mar 21 09:35:43 crc kubenswrapper[4696]: I0321 09:35:43.539313 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-kz4rv_719c50b9-02ed-4138-8ce6-624b718e5d00/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:35:43 crc kubenswrapper[4696]: I0321 09:35:43.631137 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-proc-0_e0413efc-468a-4065-9b3f-6355ec9d0e1e/cloudkitty-proc/0.log" Mar 21 09:35:44 crc kubenswrapper[4696]: I0321 09:35:44.185443 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-85f64749dc-tqgff_c9fe9d98-b02d-4fdd-918e-c296d52f7efd/init/0.log" Mar 21 09:35:44 crc kubenswrapper[4696]: I0321 09:35:44.260592 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-6k46h_7f2cc308-7143-4847-9e6d-a96e9ddd6c51/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:35:44 crc kubenswrapper[4696]: I0321 09:35:44.398928 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-85f64749dc-tqgff_c9fe9d98-b02d-4fdd-918e-c296d52f7efd/init/0.log" Mar 21 09:35:44 crc kubenswrapper[4696]: I0321 09:35:44.456024 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-85f64749dc-tqgff_c9fe9d98-b02d-4fdd-918e-c296d52f7efd/dnsmasq-dns/0.log" Mar 21 09:35:44 crc kubenswrapper[4696]: I0321 09:35:44.508964 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-zncdp_c756d18c-c925-44bd-911f-0b82d961c66e/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:35:44 crc kubenswrapper[4696]: I0321 09:35:44.627885 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_325e7183-8d8e-4a92-ab49-3dc4a043c21c/glance-httpd/0.log" Mar 21 09:35:44 crc kubenswrapper[4696]: I0321 09:35:44.672023 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_325e7183-8d8e-4a92-ab49-3dc4a043c21c/glance-log/0.log" Mar 21 09:35:44 crc kubenswrapper[4696]: I0321 09:35:44.684647 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7/glance-httpd/0.log" Mar 21 09:35:44 crc kubenswrapper[4696]: I0321 09:35:44.751234 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a2b2e727-25e3-4a1e-b688-5c5adfd9b1b7/glance-log/0.log" Mar 21 09:35:44 crc kubenswrapper[4696]: I0321 09:35:44.894959 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-6rqx2_053326a0-54a2-4838-b024-4650e216f804/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:35:45 crc kubenswrapper[4696]: I0321 09:35:45.100376 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29568061-9qzjv_93e69ad1-eb81-4a1a-ae60-6feff0891ced/keystone-cron/0.log" Mar 21 09:35:45 crc kubenswrapper[4696]: I0321 09:35:45.156772 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-85v2z_3ec7503b-5ab5-4b4e-bbea-0c64e8c1cecd/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:35:45 crc kubenswrapper[4696]: I0321 09:35:45.368008 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-696f547b5d-r9gcd_31091b53-3bf3-4704-96bc-c5c33eb55218/keystone-api/0.log" Mar 21 09:35:45 crc kubenswrapper[4696]: I0321 09:35:45.792963 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_53c78963-a90b-40f0-b223-0c70917cee6e/kube-state-metrics/0.log" Mar 21 09:35:46 crc kubenswrapper[4696]: I0321 09:35:46.211343 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-69b5b45b6f-7lhn7_1157f1b6-6891-43eb-93fc-e8e66691359a/neutron-httpd/0.log" Mar 21 09:35:46 crc kubenswrapper[4696]: I0321 09:35:46.275789 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-69b5b45b6f-7lhn7_1157f1b6-6891-43eb-93fc-e8e66691359a/neutron-api/0.log" Mar 21 09:35:46 crc kubenswrapper[4696]: I0321 09:35:46.283682 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-wkl4d_50e8cc4d-e8bf-4f1a-9c84-f40af2392cde/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:35:46 crc kubenswrapper[4696]: I0321 09:35:46.505053 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-nwqqv_76b63ef6-105a-4d99-a8bc-a8758970ffff/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:35:46 crc kubenswrapper[4696]: I0321 09:35:46.878972 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_348bda58-214e-41e3-bc04-2fdcf0e79142/nova-api-log/0.log" Mar 21 09:35:46 crc kubenswrapper[4696]: I0321 09:35:46.919834 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_7ba86395-973a-4d8b-976f-673403ae58e5/nova-cell0-conductor-conductor/0.log" Mar 21 09:35:47 crc kubenswrapper[4696]: I0321 09:35:47.270666 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_58d8d1a2-0ecf-45d2-aa4d-33a54d0ae267/nova-cell1-conductor-conductor/0.log" Mar 21 09:35:47 crc kubenswrapper[4696]: I0321 09:35:47.359220 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_f424006b-438e-46cc-8d5f-70d8d0346f40/nova-cell1-novncproxy-novncproxy/0.log" Mar 21 09:35:47 crc kubenswrapper[4696]: I0321 09:35:47.390714 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_348bda58-214e-41e3-bc04-2fdcf0e79142/nova-api-api/0.log" Mar 21 09:35:47 crc kubenswrapper[4696]: I0321 09:35:47.717249 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f812f449-0a51-4414-ac6f-31ecf7648686/nova-metadata-log/0.log" Mar 21 09:35:48 crc kubenswrapper[4696]: I0321 09:35:48.097417 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-dpstl_14ded188-15ec-4d0c-a462-d1160dfb5b47/nova-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:35:48 crc kubenswrapper[4696]: I0321 09:35:48.134630 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f812f449-0a51-4414-ac6f-31ecf7648686/nova-metadata-metadata/0.log" Mar 21 09:35:48 crc kubenswrapper[4696]: I0321 09:35:48.137209 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_8b1800dd-28c2-4614-be8a-06afc002a85e/nova-scheduler-scheduler/0.log" Mar 21 09:35:48 crc kubenswrapper[4696]: I0321 09:35:48.257715 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_57e750dc-d19c-45d4-9905-b0a15a6a642d/mysql-bootstrap/0.log" Mar 21 09:35:48 crc kubenswrapper[4696]: I0321 09:35:48.489888 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_57e750dc-d19c-45d4-9905-b0a15a6a642d/mysql-bootstrap/0.log" Mar 21 09:35:48 crc kubenswrapper[4696]: I0321 09:35:48.532692 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_57e750dc-d19c-45d4-9905-b0a15a6a642d/galera/0.log" Mar 21 09:35:48 crc kubenswrapper[4696]: I0321 09:35:48.634144 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_919a11c7-cf69-4660-8de6-6b1a54b63f26/mysql-bootstrap/0.log" Mar 21 09:35:48 crc kubenswrapper[4696]: I0321 09:35:48.751546 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_919a11c7-cf69-4660-8de6-6b1a54b63f26/galera/0.log" Mar 21 09:35:48 crc kubenswrapper[4696]: I0321 09:35:48.786454 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_919a11c7-cf69-4660-8de6-6b1a54b63f26/mysql-bootstrap/0.log" Mar 21 09:35:48 crc kubenswrapper[4696]: I0321 09:35:48.905177 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_383967e9-132f-4355-abbb-5904830137b7/openstackclient/0.log" Mar 21 09:35:48 crc kubenswrapper[4696]: I0321 09:35:48.992183 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-4vd4t_79ac7ba5-f3a1-4c5a-ab86-8517fbe0aa3a/ovn-controller/0.log" Mar 21 09:35:49 crc kubenswrapper[4696]: I0321 09:35:49.110294 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-8nzqr_63645afc-9440-40b7-ae49-9203c9dcb3f8/openstack-network-exporter/0.log" Mar 21 09:35:49 crc kubenswrapper[4696]: I0321 09:35:49.249382 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dmbpm_7f0d9af4-d6e3-4168-b83d-5a482fb51d44/ovsdb-server-init/0.log" Mar 21 09:35:49 crc kubenswrapper[4696]: I0321 09:35:49.420839 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dmbpm_7f0d9af4-d6e3-4168-b83d-5a482fb51d44/ovsdb-server-init/0.log" Mar 21 09:35:49 crc kubenswrapper[4696]: I0321 09:35:49.471924 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dmbpm_7f0d9af4-d6e3-4168-b83d-5a482fb51d44/ovs-vswitchd/0.log" Mar 21 09:35:49 crc kubenswrapper[4696]: I0321 09:35:49.539732 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dmbpm_7f0d9af4-d6e3-4168-b83d-5a482fb51d44/ovsdb-server/0.log" Mar 21 09:35:49 crc kubenswrapper[4696]: I0321 09:35:49.700413 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_dc5349db-8b6f-41a8-b3c7-d42920c3bf58/openstack-network-exporter/0.log" Mar 21 09:35:49 crc kubenswrapper[4696]: I0321 09:35:49.750714 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_dc5349db-8b6f-41a8-b3c7-d42920c3bf58/ovn-northd/0.log" Mar 21 09:35:49 crc kubenswrapper[4696]: I0321 09:35:49.790193 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-v5hcz_2c75b9c9-8d94-4541-8858-0c56d8ecb400/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:35:49 crc kubenswrapper[4696]: I0321 09:35:49.965109 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21/ovsdbserver-nb/0.log" Mar 21 09:35:49 crc kubenswrapper[4696]: I0321 09:35:49.973830 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_2e9e1eae-a6a4-4a7b-803f-c88b04b6bf21/openstack-network-exporter/0.log" Mar 21 09:35:50 crc kubenswrapper[4696]: I0321 09:35:50.128693 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_b9782617-ee6a-4ab5-aa9c-22dc382f654e/openstack-network-exporter/0.log" Mar 21 09:35:50 crc kubenswrapper[4696]: I0321 09:35:50.187239 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_b9782617-ee6a-4ab5-aa9c-22dc382f654e/ovsdbserver-sb/0.log" Mar 21 09:35:50 crc kubenswrapper[4696]: I0321 09:35:50.310412 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-56b7788dc4-f928g_3bd18ca4-e084-4fe1-85e3-7319ddd703ff/placement-api/0.log" Mar 21 09:35:50 crc kubenswrapper[4696]: I0321 09:35:50.420934 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-56b7788dc4-f928g_3bd18ca4-e084-4fe1-85e3-7319ddd703ff/placement-log/0.log" Mar 21 09:35:50 crc kubenswrapper[4696]: I0321 09:35:50.459993 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_c2c63b45-de34-43a6-bb8b-aed4b00922c8/init-config-reloader/0.log" Mar 21 09:35:50 crc kubenswrapper[4696]: I0321 09:35:50.602003 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_c2c63b45-de34-43a6-bb8b-aed4b00922c8/config-reloader/0.log" Mar 21 09:35:50 crc kubenswrapper[4696]: I0321 09:35:50.617778 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_c2c63b45-de34-43a6-bb8b-aed4b00922c8/init-config-reloader/0.log" Mar 21 09:35:50 crc kubenswrapper[4696]: I0321 09:35:50.680222 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_c2c63b45-de34-43a6-bb8b-aed4b00922c8/thanos-sidecar/0.log" Mar 21 09:35:50 crc kubenswrapper[4696]: I0321 09:35:50.688425 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_c2c63b45-de34-43a6-bb8b-aed4b00922c8/prometheus/0.log" Mar 21 09:35:50 crc kubenswrapper[4696]: I0321 09:35:50.851516 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0f40066c-3b43-42a5-9bb2-c8fb94a921db/setup-container/0.log" Mar 21 09:35:51 crc kubenswrapper[4696]: I0321 09:35:51.013228 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0f40066c-3b43-42a5-9bb2-c8fb94a921db/setup-container/0.log" Mar 21 09:35:51 crc kubenswrapper[4696]: I0321 09:35:51.030263 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0f40066c-3b43-42a5-9bb2-c8fb94a921db/rabbitmq/0.log" Mar 21 09:35:51 crc kubenswrapper[4696]: I0321 09:35:51.080133 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_56d6eab2-0e95-402a-afb7-2707ca41b144/setup-container/0.log" Mar 21 09:35:51 crc kubenswrapper[4696]: I0321 09:35:51.404496 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-5rb2h_e19a30a1-9816-4c71-9803-ed0dbbdad615/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:35:51 crc kubenswrapper[4696]: I0321 09:35:51.413560 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_56d6eab2-0e95-402a-afb7-2707ca41b144/rabbitmq/0.log" Mar 21 09:35:51 crc kubenswrapper[4696]: I0321 09:35:51.457052 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_56d6eab2-0e95-402a-afb7-2707ca41b144/setup-container/0.log" Mar 21 09:35:51 crc kubenswrapper[4696]: I0321 09:35:51.787296 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-kjr8n_93d9988c-26af-48d1-b43e-7c897cad7235/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:35:51 crc kubenswrapper[4696]: I0321 09:35:51.806795 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-6cdss_480a9890-311e-44b9-bfac-4c375e7cdf1e/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:35:52 crc kubenswrapper[4696]: I0321 09:35:52.012510 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-km5wx_5a6914f4-0b0c-4f4a-8e09-b9b8aeb3dd6f/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:35:52 crc kubenswrapper[4696]: I0321 09:35:52.084586 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-ckxp8_4f9d8f9c-fde1-4198-b24e-bcffe8ace9b6/ssh-known-hosts-edpm-deployment/0.log" Mar 21 09:35:52 crc kubenswrapper[4696]: I0321 09:35:52.386096 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7b65745bbc-5b5zz_4abc84ca-2aab-4423-b934-c61f3ef8ea6d/proxy-server/0.log" Mar 21 09:35:52 crc kubenswrapper[4696]: I0321 09:35:52.412795 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7b65745bbc-5b5zz_4abc84ca-2aab-4423-b934-c61f3ef8ea6d/proxy-httpd/0.log" Mar 21 09:35:52 crc kubenswrapper[4696]: I0321 09:35:52.535020 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-jznlt_efd827ca-ae90-4cca-a77d-72fce76c47aa/swift-ring-rebalance/0.log" Mar 21 09:35:52 crc kubenswrapper[4696]: I0321 09:35:52.606098 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/account-auditor/0.log" Mar 21 09:35:52 crc kubenswrapper[4696]: I0321 09:35:52.655290 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/account-reaper/0.log" Mar 21 09:35:52 crc kubenswrapper[4696]: I0321 09:35:52.757883 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/account-replicator/0.log" Mar 21 09:35:53 crc kubenswrapper[4696]: I0321 09:35:53.334292 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/container-auditor/0.log" Mar 21 09:35:53 crc kubenswrapper[4696]: I0321 09:35:53.376321 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/account-server/0.log" Mar 21 09:35:53 crc kubenswrapper[4696]: I0321 09:35:53.413429 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/container-server/0.log" Mar 21 09:35:53 crc kubenswrapper[4696]: I0321 09:35:53.413717 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/container-replicator/0.log" Mar 21 09:35:53 crc kubenswrapper[4696]: I0321 09:35:53.548181 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/container-updater/0.log" Mar 21 09:35:53 crc kubenswrapper[4696]: I0321 09:35:53.614325 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/object-auditor/0.log" Mar 21 09:35:53 crc kubenswrapper[4696]: I0321 09:35:53.649012 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/object-expirer/0.log" Mar 21 09:35:53 crc kubenswrapper[4696]: I0321 09:35:53.679401 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/object-replicator/0.log" Mar 21 09:35:53 crc kubenswrapper[4696]: I0321 09:35:53.746123 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/object-server/0.log" Mar 21 09:35:53 crc kubenswrapper[4696]: I0321 09:35:53.861479 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/object-updater/0.log" Mar 21 09:35:53 crc kubenswrapper[4696]: I0321 09:35:53.884600 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/rsync/0.log" Mar 21 09:35:53 crc kubenswrapper[4696]: I0321 09:35:53.905580 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1088ee04-73e8-4aba-aad4-1cd7d26e20aa/swift-recon-cron/0.log" Mar 21 09:35:54 crc kubenswrapper[4696]: I0321 09:35:54.180006 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_1a6c65a7-b20f-4675-a6fc-c4e1988a2a1e/tempest-tests-tempest-tests-runner/0.log" Mar 21 09:35:54 crc kubenswrapper[4696]: I0321 09:35:54.346815 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_e6c3f461-61e5-443c-911d-0b8a1ba72a96/test-operator-logs-container/0.log" Mar 21 09:35:54 crc kubenswrapper[4696]: I0321 09:35:54.493924 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-z8mr8_2b3f6677-db77-456b-aff2-9deee1fe26f9/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:35:54 crc kubenswrapper[4696]: I0321 09:35:54.924545 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-5ctj8_a8cdd5da-1ba0-4034-8a15-e3e86352772f/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 21 09:35:58 crc kubenswrapper[4696]: I0321 09:35:58.340634 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_dd0d5d6d-801d-4446-a325-f82b2d55bca6/memcached/0.log" Mar 21 09:36:00 crc kubenswrapper[4696]: I0321 09:36:00.138408 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568096-hq7ss"] Mar 21 09:36:00 crc kubenswrapper[4696]: E0321 09:36:00.139079 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eceb525-7807-4676-8da7-0f7b74ce478c" containerName="container-00" Mar 21 09:36:00 crc kubenswrapper[4696]: I0321 09:36:00.139092 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eceb525-7807-4676-8da7-0f7b74ce478c" containerName="container-00" Mar 21 09:36:00 crc kubenswrapper[4696]: I0321 09:36:00.139272 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eceb525-7807-4676-8da7-0f7b74ce478c" containerName="container-00" Mar 21 09:36:00 crc kubenswrapper[4696]: I0321 09:36:00.140038 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568096-hq7ss" Mar 21 09:36:00 crc kubenswrapper[4696]: I0321 09:36:00.142040 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:36:00 crc kubenswrapper[4696]: I0321 09:36:00.142665 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:36:00 crc kubenswrapper[4696]: I0321 09:36:00.142777 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:36:00 crc kubenswrapper[4696]: I0321 09:36:00.156742 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568096-hq7ss"] Mar 21 09:36:00 crc kubenswrapper[4696]: I0321 09:36:00.269270 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67z76\" (UniqueName: \"kubernetes.io/projected/73788553-329c-410c-bf8a-549ff702848d-kube-api-access-67z76\") pod \"auto-csr-approver-29568096-hq7ss\" (UID: \"73788553-329c-410c-bf8a-549ff702848d\") " pod="openshift-infra/auto-csr-approver-29568096-hq7ss" Mar 21 09:36:00 crc kubenswrapper[4696]: I0321 09:36:00.372152 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67z76\" (UniqueName: \"kubernetes.io/projected/73788553-329c-410c-bf8a-549ff702848d-kube-api-access-67z76\") pod \"auto-csr-approver-29568096-hq7ss\" (UID: \"73788553-329c-410c-bf8a-549ff702848d\") " pod="openshift-infra/auto-csr-approver-29568096-hq7ss" Mar 21 09:36:00 crc kubenswrapper[4696]: I0321 09:36:00.405835 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67z76\" (UniqueName: \"kubernetes.io/projected/73788553-329c-410c-bf8a-549ff702848d-kube-api-access-67z76\") pod \"auto-csr-approver-29568096-hq7ss\" (UID: \"73788553-329c-410c-bf8a-549ff702848d\") " pod="openshift-infra/auto-csr-approver-29568096-hq7ss" Mar 21 09:36:00 crc kubenswrapper[4696]: I0321 09:36:00.458418 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568096-hq7ss" Mar 21 09:36:00 crc kubenswrapper[4696]: I0321 09:36:00.947126 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568096-hq7ss"] Mar 21 09:36:00 crc kubenswrapper[4696]: I0321 09:36:00.950353 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 09:36:01 crc kubenswrapper[4696]: I0321 09:36:01.825874 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568096-hq7ss" event={"ID":"73788553-329c-410c-bf8a-549ff702848d","Type":"ContainerStarted","Data":"42752929f3d050f5ab69d7bdef59e9d5093b583c137acdcd26defd9f47ebc989"} Mar 21 09:36:02 crc kubenswrapper[4696]: I0321 09:36:02.835749 4696 generic.go:334] "Generic (PLEG): container finished" podID="73788553-329c-410c-bf8a-549ff702848d" containerID="f168e3e0b46f425a9f90203de70925dc5494348a1f153d0ad4a5d9f5eb135eb2" exitCode=0 Mar 21 09:36:02 crc kubenswrapper[4696]: I0321 09:36:02.835847 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568096-hq7ss" event={"ID":"73788553-329c-410c-bf8a-549ff702848d","Type":"ContainerDied","Data":"f168e3e0b46f425a9f90203de70925dc5494348a1f153d0ad4a5d9f5eb135eb2"} Mar 21 09:36:04 crc kubenswrapper[4696]: I0321 09:36:04.716166 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568096-hq7ss" Mar 21 09:36:04 crc kubenswrapper[4696]: I0321 09:36:04.858097 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67z76\" (UniqueName: \"kubernetes.io/projected/73788553-329c-410c-bf8a-549ff702848d-kube-api-access-67z76\") pod \"73788553-329c-410c-bf8a-549ff702848d\" (UID: \"73788553-329c-410c-bf8a-549ff702848d\") " Mar 21 09:36:04 crc kubenswrapper[4696]: I0321 09:36:04.876152 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73788553-329c-410c-bf8a-549ff702848d-kube-api-access-67z76" (OuterVolumeSpecName: "kube-api-access-67z76") pod "73788553-329c-410c-bf8a-549ff702848d" (UID: "73788553-329c-410c-bf8a-549ff702848d"). InnerVolumeSpecName "kube-api-access-67z76". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:36:04 crc kubenswrapper[4696]: I0321 09:36:04.878562 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568096-hq7ss" event={"ID":"73788553-329c-410c-bf8a-549ff702848d","Type":"ContainerDied","Data":"42752929f3d050f5ab69d7bdef59e9d5093b583c137acdcd26defd9f47ebc989"} Mar 21 09:36:04 crc kubenswrapper[4696]: I0321 09:36:04.878669 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42752929f3d050f5ab69d7bdef59e9d5093b583c137acdcd26defd9f47ebc989" Mar 21 09:36:04 crc kubenswrapper[4696]: I0321 09:36:04.878807 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568096-hq7ss" Mar 21 09:36:04 crc kubenswrapper[4696]: I0321 09:36:04.961293 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67z76\" (UniqueName: \"kubernetes.io/projected/73788553-329c-410c-bf8a-549ff702848d-kube-api-access-67z76\") on node \"crc\" DevicePath \"\"" Mar 21 09:36:05 crc kubenswrapper[4696]: I0321 09:36:05.791861 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568090-g4t47"] Mar 21 09:36:05 crc kubenswrapper[4696]: I0321 09:36:05.803702 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568090-g4t47"] Mar 21 09:36:06 crc kubenswrapper[4696]: I0321 09:36:06.545663 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb1746c4-d2da-45f7-90a2-d28635623d57" path="/var/lib/kubelet/pods/bb1746c4-d2da-45f7-90a2-d28635623d57/volumes" Mar 21 09:36:11 crc kubenswrapper[4696]: I0321 09:36:11.373083 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mv2td"] Mar 21 09:36:11 crc kubenswrapper[4696]: E0321 09:36:11.374340 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73788553-329c-410c-bf8a-549ff702848d" containerName="oc" Mar 21 09:36:11 crc kubenswrapper[4696]: I0321 09:36:11.374354 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="73788553-329c-410c-bf8a-549ff702848d" containerName="oc" Mar 21 09:36:11 crc kubenswrapper[4696]: I0321 09:36:11.374578 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="73788553-329c-410c-bf8a-549ff702848d" containerName="oc" Mar 21 09:36:11 crc kubenswrapper[4696]: I0321 09:36:11.376361 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mv2td" Mar 21 09:36:11 crc kubenswrapper[4696]: I0321 09:36:11.387953 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mv2td"] Mar 21 09:36:11 crc kubenswrapper[4696]: I0321 09:36:11.505730 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38f08644-f314-41b3-ab47-0f32a8de114c-catalog-content\") pod \"redhat-operators-mv2td\" (UID: \"38f08644-f314-41b3-ab47-0f32a8de114c\") " pod="openshift-marketplace/redhat-operators-mv2td" Mar 21 09:36:11 crc kubenswrapper[4696]: I0321 09:36:11.505799 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjjrs\" (UniqueName: \"kubernetes.io/projected/38f08644-f314-41b3-ab47-0f32a8de114c-kube-api-access-tjjrs\") pod \"redhat-operators-mv2td\" (UID: \"38f08644-f314-41b3-ab47-0f32a8de114c\") " pod="openshift-marketplace/redhat-operators-mv2td" Mar 21 09:36:11 crc kubenswrapper[4696]: I0321 09:36:11.506544 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38f08644-f314-41b3-ab47-0f32a8de114c-utilities\") pod \"redhat-operators-mv2td\" (UID: \"38f08644-f314-41b3-ab47-0f32a8de114c\") " pod="openshift-marketplace/redhat-operators-mv2td" Mar 21 09:36:11 crc kubenswrapper[4696]: I0321 09:36:11.608335 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38f08644-f314-41b3-ab47-0f32a8de114c-utilities\") pod \"redhat-operators-mv2td\" (UID: \"38f08644-f314-41b3-ab47-0f32a8de114c\") " pod="openshift-marketplace/redhat-operators-mv2td" Mar 21 09:36:11 crc kubenswrapper[4696]: I0321 09:36:11.609223 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38f08644-f314-41b3-ab47-0f32a8de114c-catalog-content\") pod \"redhat-operators-mv2td\" (UID: \"38f08644-f314-41b3-ab47-0f32a8de114c\") " pod="openshift-marketplace/redhat-operators-mv2td" Mar 21 09:36:11 crc kubenswrapper[4696]: I0321 09:36:11.609284 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjjrs\" (UniqueName: \"kubernetes.io/projected/38f08644-f314-41b3-ab47-0f32a8de114c-kube-api-access-tjjrs\") pod \"redhat-operators-mv2td\" (UID: \"38f08644-f314-41b3-ab47-0f32a8de114c\") " pod="openshift-marketplace/redhat-operators-mv2td" Mar 21 09:36:11 crc kubenswrapper[4696]: I0321 09:36:11.609680 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38f08644-f314-41b3-ab47-0f32a8de114c-utilities\") pod \"redhat-operators-mv2td\" (UID: \"38f08644-f314-41b3-ab47-0f32a8de114c\") " pod="openshift-marketplace/redhat-operators-mv2td" Mar 21 09:36:11 crc kubenswrapper[4696]: I0321 09:36:11.609746 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38f08644-f314-41b3-ab47-0f32a8de114c-catalog-content\") pod \"redhat-operators-mv2td\" (UID: \"38f08644-f314-41b3-ab47-0f32a8de114c\") " pod="openshift-marketplace/redhat-operators-mv2td" Mar 21 09:36:12 crc kubenswrapper[4696]: I0321 09:36:12.169425 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjjrs\" (UniqueName: \"kubernetes.io/projected/38f08644-f314-41b3-ab47-0f32a8de114c-kube-api-access-tjjrs\") pod \"redhat-operators-mv2td\" (UID: \"38f08644-f314-41b3-ab47-0f32a8de114c\") " pod="openshift-marketplace/redhat-operators-mv2td" Mar 21 09:36:12 crc kubenswrapper[4696]: I0321 09:36:12.305156 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mv2td" Mar 21 09:36:12 crc kubenswrapper[4696]: I0321 09:36:12.839663 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mv2td"] Mar 21 09:36:12 crc kubenswrapper[4696]: I0321 09:36:12.976986 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mv2td" event={"ID":"38f08644-f314-41b3-ab47-0f32a8de114c","Type":"ContainerStarted","Data":"d9ab7449bb8697d10320e50a0b64bfbfca0aa316b50a2f0d48cc7a7b08abe3f0"} Mar 21 09:36:13 crc kubenswrapper[4696]: I0321 09:36:13.987520 4696 generic.go:334] "Generic (PLEG): container finished" podID="38f08644-f314-41b3-ab47-0f32a8de114c" containerID="e296e0bdf8dfbebdc85a39d5b446717852029b459d583f180d0896b9e656bac6" exitCode=0 Mar 21 09:36:13 crc kubenswrapper[4696]: I0321 09:36:13.987606 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mv2td" event={"ID":"38f08644-f314-41b3-ab47-0f32a8de114c","Type":"ContainerDied","Data":"e296e0bdf8dfbebdc85a39d5b446717852029b459d583f180d0896b9e656bac6"} Mar 21 09:36:14 crc kubenswrapper[4696]: I0321 09:36:14.997238 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mv2td" event={"ID":"38f08644-f314-41b3-ab47-0f32a8de114c","Type":"ContainerStarted","Data":"d8c679c3c32cc094ceadb07c5e7681b246dbd7e90db90a3a010e82282e558eaf"} Mar 21 09:36:18 crc kubenswrapper[4696]: I0321 09:36:18.024759 4696 generic.go:334] "Generic (PLEG): container finished" podID="38f08644-f314-41b3-ab47-0f32a8de114c" containerID="d8c679c3c32cc094ceadb07c5e7681b246dbd7e90db90a3a010e82282e558eaf" exitCode=0 Mar 21 09:36:18 crc kubenswrapper[4696]: I0321 09:36:18.024835 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mv2td" event={"ID":"38f08644-f314-41b3-ab47-0f32a8de114c","Type":"ContainerDied","Data":"d8c679c3c32cc094ceadb07c5e7681b246dbd7e90db90a3a010e82282e558eaf"} Mar 21 09:36:19 crc kubenswrapper[4696]: I0321 09:36:19.036679 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mv2td" event={"ID":"38f08644-f314-41b3-ab47-0f32a8de114c","Type":"ContainerStarted","Data":"8676b9b2388f035c4c7cb164477a93fb022d493d75901085dedc3fa419614a1e"} Mar 21 09:36:19 crc kubenswrapper[4696]: I0321 09:36:19.062494 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mv2td" podStartSLOduration=3.627476772 podStartE2EDuration="8.062474829s" podCreationTimestamp="2026-03-21 09:36:11 +0000 UTC" firstStartedPulling="2026-03-21 09:36:13.989312271 +0000 UTC m=+4108.110192984" lastFinishedPulling="2026-03-21 09:36:18.424310328 +0000 UTC m=+4112.545191041" observedRunningTime="2026-03-21 09:36:19.054270565 +0000 UTC m=+4113.175151298" watchObservedRunningTime="2026-03-21 09:36:19.062474829 +0000 UTC m=+4113.183355542" Mar 21 09:36:22 crc kubenswrapper[4696]: I0321 09:36:22.250318 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t_01e5ce30-73e2-4e41-9d25-6ac3c67f9361/util/0.log" Mar 21 09:36:22 crc kubenswrapper[4696]: I0321 09:36:22.306311 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mv2td" Mar 21 09:36:22 crc kubenswrapper[4696]: I0321 09:36:22.306412 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mv2td" Mar 21 09:36:22 crc kubenswrapper[4696]: I0321 09:36:22.513579 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t_01e5ce30-73e2-4e41-9d25-6ac3c67f9361/util/0.log" Mar 21 09:36:22 crc kubenswrapper[4696]: I0321 09:36:22.543943 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t_01e5ce30-73e2-4e41-9d25-6ac3c67f9361/pull/0.log" Mar 21 09:36:22 crc kubenswrapper[4696]: I0321 09:36:22.566493 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t_01e5ce30-73e2-4e41-9d25-6ac3c67f9361/pull/0.log" Mar 21 09:36:22 crc kubenswrapper[4696]: I0321 09:36:22.691592 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t_01e5ce30-73e2-4e41-9d25-6ac3c67f9361/util/0.log" Mar 21 09:36:22 crc kubenswrapper[4696]: I0321 09:36:22.737459 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t_01e5ce30-73e2-4e41-9d25-6ac3c67f9361/extract/0.log" Mar 21 09:36:22 crc kubenswrapper[4696]: I0321 09:36:22.774118 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_364d18c58bdc83c0cd1e7a53ac7c4186d3fb300bc088e4665ae81f1d8bflz6t_01e5ce30-73e2-4e41-9d25-6ac3c67f9361/pull/0.log" Mar 21 09:36:23 crc kubenswrapper[4696]: I0321 09:36:23.272018 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6cc65c69fc-m8ffd_530eca0b-da29-4883-ad3a-1faa1563c65d/manager/0.log" Mar 21 09:36:23 crc kubenswrapper[4696]: I0321 09:36:23.357949 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mv2td" podUID="38f08644-f314-41b3-ab47-0f32a8de114c" containerName="registry-server" probeResult="failure" output=< Mar 21 09:36:23 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Mar 21 09:36:23 crc kubenswrapper[4696]: > Mar 21 09:36:23 crc kubenswrapper[4696]: I0321 09:36:23.645130 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7d559dcdbd-zxgmc_9ce080a1-65e5-4235-b181-448e8987c590/manager/0.log" Mar 21 09:36:24 crc kubenswrapper[4696]: I0321 09:36:24.113506 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-66dd9d474d-qf5kd_58ec1c2d-43a6-4582-9a21-f1083777ffe5/manager/0.log" Mar 21 09:36:24 crc kubenswrapper[4696]: I0321 09:36:24.343767 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-64dc66d669-8685h_9b05c8a9-af90-4497-94f3-8fed890ab3b0/manager/0.log" Mar 21 09:36:24 crc kubenswrapper[4696]: I0321 09:36:24.976941 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5595c7d6ff-7cg24_0048e137-52d0-476b-845a-4c42829de1e2/manager/0.log" Mar 21 09:36:25 crc kubenswrapper[4696]: I0321 09:36:25.003325 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6b77b7676d-vhnjl_74fc6c6a-6eb4-4459-aed9-62073e0472e4/manager/0.log" Mar 21 09:36:25 crc kubenswrapper[4696]: I0321 09:36:25.325493 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-76b87776c9-rcz2g_c31077f0-de38-4de7-8421-a5ea7ac3ab84/manager/0.log" Mar 21 09:36:25 crc kubenswrapper[4696]: I0321 09:36:25.413422 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6d77645966-qsd27_3cd62d0b-d740-4bef-ac3c-993b5ee05d40/manager/0.log" Mar 21 09:36:25 crc kubenswrapper[4696]: I0321 09:36:25.563251 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-fbf7bbb96-745cq_8e08ed5d-23b6-4066-a86f-3019f8cb34c8/manager/0.log" Mar 21 09:36:25 crc kubenswrapper[4696]: I0321 09:36:25.716921 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6f5b7bcd4-pzgkp_a4bbdaf9-e480-4cd1-9bf8-304cf5266949/manager/0.log" Mar 21 09:36:25 crc kubenswrapper[4696]: I0321 09:36:25.978243 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6744dd545c-mvqjq_f9ca1637-123c-47eb-8183-5394a0ba6f89/manager/0.log" Mar 21 09:36:26 crc kubenswrapper[4696]: I0321 09:36:26.142617 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-bc5c78db9-c9skl_50f56ad0-1409-4619-9b7a-887ba6997874/manager/0.log" Mar 21 09:36:26 crc kubenswrapper[4696]: I0321 09:36:26.197757 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-56f74467c6-fzv5q_828a9554-fa0a-4458-9d63-b9e6bd6bde49/manager/0.log" Mar 21 09:36:26 crc kubenswrapper[4696]: I0321 09:36:26.401024 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-c5677dc5d-jpc89_026783e3-cff7-40bf-9f26-2d5e8b6006e7/manager/0.log" Mar 21 09:36:26 crc kubenswrapper[4696]: I0321 09:36:26.965882 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-k24x8_2e3b5f4d-52bd-4d2c-b629-74fc14699179/registry-server/0.log" Mar 21 09:36:27 crc kubenswrapper[4696]: I0321 09:36:27.060043 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-888f66487-rsc92_bfd4f311-01db-41ce-8357-93860d30b442/operator/0.log" Mar 21 09:36:27 crc kubenswrapper[4696]: I0321 09:36:27.264051 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-846c4cdcb7-4s5br_1c77ab60-1377-4880-80a9-eb1af2009bca/manager/0.log" Mar 21 09:36:27 crc kubenswrapper[4696]: I0321 09:36:27.438231 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-659fb58c6b-m5bx9_d2672ea4-e1e0-4809-b4e9-04e2cece2179/manager/0.log" Mar 21 09:36:27 crc kubenswrapper[4696]: I0321 09:36:27.566409 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-kvljx_1ff5cfb7-37ad-4f6a-84f1-e8dbdef06084/operator/0.log" Mar 21 09:36:27 crc kubenswrapper[4696]: I0321 09:36:27.773097 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-867f54bc44-n94f2_5fe44a3f-65e2-481b-9d1d-f3b3c24944db/manager/0.log" Mar 21 09:36:28 crc kubenswrapper[4696]: I0321 09:36:28.061615 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-8467ccb4c8-zk58x_46974657-c404-4081-a4ef-5fc99237a02f/manager/0.log" Mar 21 09:36:28 crc kubenswrapper[4696]: I0321 09:36:28.414510 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-74d6f7b5c-5x9b6_f593acbc-0a73-41ed-a949-02a3556dfc70/manager/0.log" Mar 21 09:36:28 crc kubenswrapper[4696]: I0321 09:36:28.745118 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-69d78c9949-5whvx_0ecf9ee3-d5c9-4677-b264-0ddae28df715/manager/0.log" Mar 21 09:36:28 crc kubenswrapper[4696]: I0321 09:36:28.810921 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-b7fdfc6b-hrhxc_fbbc65c6-7696-4795-8b4b-7d9c7e315bf0/manager/0.log" Mar 21 09:36:33 crc kubenswrapper[4696]: I0321 09:36:33.358944 4696 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mv2td" podUID="38f08644-f314-41b3-ab47-0f32a8de114c" containerName="registry-server" probeResult="failure" output=< Mar 21 09:36:33 crc kubenswrapper[4696]: timeout: failed to connect service ":50051" within 1s Mar 21 09:36:33 crc kubenswrapper[4696]: > Mar 21 09:36:33 crc kubenswrapper[4696]: I0321 09:36:33.745782 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5cfd84c587-ldj9m_53a9e616-007d-4c0c-8b0a-799d7188d9ab/manager/0.log" Mar 21 09:36:42 crc kubenswrapper[4696]: I0321 09:36:42.354394 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mv2td" Mar 21 09:36:42 crc kubenswrapper[4696]: I0321 09:36:42.406086 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mv2td" Mar 21 09:36:42 crc kubenswrapper[4696]: I0321 09:36:42.589376 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mv2td"] Mar 21 09:36:44 crc kubenswrapper[4696]: I0321 09:36:44.318385 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mv2td" podUID="38f08644-f314-41b3-ab47-0f32a8de114c" containerName="registry-server" containerID="cri-o://8676b9b2388f035c4c7cb164477a93fb022d493d75901085dedc3fa419614a1e" gracePeriod=2 Mar 21 09:36:44 crc kubenswrapper[4696]: I0321 09:36:44.766259 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mv2td" Mar 21 09:36:44 crc kubenswrapper[4696]: I0321 09:36:44.888856 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38f08644-f314-41b3-ab47-0f32a8de114c-catalog-content\") pod \"38f08644-f314-41b3-ab47-0f32a8de114c\" (UID: \"38f08644-f314-41b3-ab47-0f32a8de114c\") " Mar 21 09:36:44 crc kubenswrapper[4696]: I0321 09:36:44.888944 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjjrs\" (UniqueName: \"kubernetes.io/projected/38f08644-f314-41b3-ab47-0f32a8de114c-kube-api-access-tjjrs\") pod \"38f08644-f314-41b3-ab47-0f32a8de114c\" (UID: \"38f08644-f314-41b3-ab47-0f32a8de114c\") " Mar 21 09:36:44 crc kubenswrapper[4696]: I0321 09:36:44.889096 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38f08644-f314-41b3-ab47-0f32a8de114c-utilities\") pod \"38f08644-f314-41b3-ab47-0f32a8de114c\" (UID: \"38f08644-f314-41b3-ab47-0f32a8de114c\") " Mar 21 09:36:44 crc kubenswrapper[4696]: I0321 09:36:44.890202 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38f08644-f314-41b3-ab47-0f32a8de114c-utilities" (OuterVolumeSpecName: "utilities") pod "38f08644-f314-41b3-ab47-0f32a8de114c" (UID: "38f08644-f314-41b3-ab47-0f32a8de114c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:36:44 crc kubenswrapper[4696]: I0321 09:36:44.895445 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38f08644-f314-41b3-ab47-0f32a8de114c-kube-api-access-tjjrs" (OuterVolumeSpecName: "kube-api-access-tjjrs") pod "38f08644-f314-41b3-ab47-0f32a8de114c" (UID: "38f08644-f314-41b3-ab47-0f32a8de114c"). InnerVolumeSpecName "kube-api-access-tjjrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:36:44 crc kubenswrapper[4696]: I0321 09:36:44.991020 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38f08644-f314-41b3-ab47-0f32a8de114c-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 09:36:44 crc kubenswrapper[4696]: I0321 09:36:44.991072 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjjrs\" (UniqueName: \"kubernetes.io/projected/38f08644-f314-41b3-ab47-0f32a8de114c-kube-api-access-tjjrs\") on node \"crc\" DevicePath \"\"" Mar 21 09:36:45 crc kubenswrapper[4696]: I0321 09:36:45.000495 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38f08644-f314-41b3-ab47-0f32a8de114c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38f08644-f314-41b3-ab47-0f32a8de114c" (UID: "38f08644-f314-41b3-ab47-0f32a8de114c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:36:45 crc kubenswrapper[4696]: I0321 09:36:45.093115 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38f08644-f314-41b3-ab47-0f32a8de114c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 09:36:45 crc kubenswrapper[4696]: I0321 09:36:45.331987 4696 generic.go:334] "Generic (PLEG): container finished" podID="38f08644-f314-41b3-ab47-0f32a8de114c" containerID="8676b9b2388f035c4c7cb164477a93fb022d493d75901085dedc3fa419614a1e" exitCode=0 Mar 21 09:36:45 crc kubenswrapper[4696]: I0321 09:36:45.332026 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mv2td" event={"ID":"38f08644-f314-41b3-ab47-0f32a8de114c","Type":"ContainerDied","Data":"8676b9b2388f035c4c7cb164477a93fb022d493d75901085dedc3fa419614a1e"} Mar 21 09:36:45 crc kubenswrapper[4696]: I0321 09:36:45.332052 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mv2td" event={"ID":"38f08644-f314-41b3-ab47-0f32a8de114c","Type":"ContainerDied","Data":"d9ab7449bb8697d10320e50a0b64bfbfca0aa316b50a2f0d48cc7a7b08abe3f0"} Mar 21 09:36:45 crc kubenswrapper[4696]: I0321 09:36:45.332071 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mv2td" Mar 21 09:36:45 crc kubenswrapper[4696]: I0321 09:36:45.332087 4696 scope.go:117] "RemoveContainer" containerID="8676b9b2388f035c4c7cb164477a93fb022d493d75901085dedc3fa419614a1e" Mar 21 09:36:45 crc kubenswrapper[4696]: I0321 09:36:45.366541 4696 scope.go:117] "RemoveContainer" containerID="d8c679c3c32cc094ceadb07c5e7681b246dbd7e90db90a3a010e82282e558eaf" Mar 21 09:36:45 crc kubenswrapper[4696]: I0321 09:36:45.367255 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mv2td"] Mar 21 09:36:45 crc kubenswrapper[4696]: I0321 09:36:45.378954 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mv2td"] Mar 21 09:36:45 crc kubenswrapper[4696]: I0321 09:36:45.393712 4696 scope.go:117] "RemoveContainer" containerID="e296e0bdf8dfbebdc85a39d5b446717852029b459d583f180d0896b9e656bac6" Mar 21 09:36:45 crc kubenswrapper[4696]: I0321 09:36:45.443408 4696 scope.go:117] "RemoveContainer" containerID="8676b9b2388f035c4c7cb164477a93fb022d493d75901085dedc3fa419614a1e" Mar 21 09:36:45 crc kubenswrapper[4696]: E0321 09:36:45.445526 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8676b9b2388f035c4c7cb164477a93fb022d493d75901085dedc3fa419614a1e\": container with ID starting with 8676b9b2388f035c4c7cb164477a93fb022d493d75901085dedc3fa419614a1e not found: ID does not exist" containerID="8676b9b2388f035c4c7cb164477a93fb022d493d75901085dedc3fa419614a1e" Mar 21 09:36:45 crc kubenswrapper[4696]: I0321 09:36:45.445566 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8676b9b2388f035c4c7cb164477a93fb022d493d75901085dedc3fa419614a1e"} err="failed to get container status \"8676b9b2388f035c4c7cb164477a93fb022d493d75901085dedc3fa419614a1e\": rpc error: code = NotFound desc = could not find container \"8676b9b2388f035c4c7cb164477a93fb022d493d75901085dedc3fa419614a1e\": container with ID starting with 8676b9b2388f035c4c7cb164477a93fb022d493d75901085dedc3fa419614a1e not found: ID does not exist" Mar 21 09:36:45 crc kubenswrapper[4696]: I0321 09:36:45.445588 4696 scope.go:117] "RemoveContainer" containerID="d8c679c3c32cc094ceadb07c5e7681b246dbd7e90db90a3a010e82282e558eaf" Mar 21 09:36:45 crc kubenswrapper[4696]: E0321 09:36:45.445867 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8c679c3c32cc094ceadb07c5e7681b246dbd7e90db90a3a010e82282e558eaf\": container with ID starting with d8c679c3c32cc094ceadb07c5e7681b246dbd7e90db90a3a010e82282e558eaf not found: ID does not exist" containerID="d8c679c3c32cc094ceadb07c5e7681b246dbd7e90db90a3a010e82282e558eaf" Mar 21 09:36:45 crc kubenswrapper[4696]: I0321 09:36:45.445895 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8c679c3c32cc094ceadb07c5e7681b246dbd7e90db90a3a010e82282e558eaf"} err="failed to get container status \"d8c679c3c32cc094ceadb07c5e7681b246dbd7e90db90a3a010e82282e558eaf\": rpc error: code = NotFound desc = could not find container \"d8c679c3c32cc094ceadb07c5e7681b246dbd7e90db90a3a010e82282e558eaf\": container with ID starting with d8c679c3c32cc094ceadb07c5e7681b246dbd7e90db90a3a010e82282e558eaf not found: ID does not exist" Mar 21 09:36:45 crc kubenswrapper[4696]: I0321 09:36:45.445914 4696 scope.go:117] "RemoveContainer" containerID="e296e0bdf8dfbebdc85a39d5b446717852029b459d583f180d0896b9e656bac6" Mar 21 09:36:45 crc kubenswrapper[4696]: E0321 09:36:45.447784 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e296e0bdf8dfbebdc85a39d5b446717852029b459d583f180d0896b9e656bac6\": container with ID starting with e296e0bdf8dfbebdc85a39d5b446717852029b459d583f180d0896b9e656bac6 not found: ID does not exist" containerID="e296e0bdf8dfbebdc85a39d5b446717852029b459d583f180d0896b9e656bac6" Mar 21 09:36:45 crc kubenswrapper[4696]: I0321 09:36:45.447851 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e296e0bdf8dfbebdc85a39d5b446717852029b459d583f180d0896b9e656bac6"} err="failed to get container status \"e296e0bdf8dfbebdc85a39d5b446717852029b459d583f180d0896b9e656bac6\": rpc error: code = NotFound desc = could not find container \"e296e0bdf8dfbebdc85a39d5b446717852029b459d583f180d0896b9e656bac6\": container with ID starting with e296e0bdf8dfbebdc85a39d5b446717852029b459d583f180d0896b9e656bac6 not found: ID does not exist" Mar 21 09:36:46 crc kubenswrapper[4696]: I0321 09:36:46.547258 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38f08644-f314-41b3-ab47-0f32a8de114c" path="/var/lib/kubelet/pods/38f08644-f314-41b3-ab47-0f32a8de114c/volumes" Mar 21 09:36:49 crc kubenswrapper[4696]: I0321 09:36:49.625194 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-5mlz9_6769f68b-1f58-48e4-8076-14fcdb9c8036/control-plane-machine-set-operator/0.log" Mar 21 09:36:49 crc kubenswrapper[4696]: I0321 09:36:49.863333 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-h5tfm_f62330e9-cdc5-469b-82ba-3364ce173686/kube-rbac-proxy/0.log" Mar 21 09:36:49 crc kubenswrapper[4696]: I0321 09:36:49.881164 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-h5tfm_f62330e9-cdc5-469b-82ba-3364ce173686/machine-api-operator/0.log" Mar 21 09:37:00 crc kubenswrapper[4696]: I0321 09:37:00.341019 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:37:00 crc kubenswrapper[4696]: I0321 09:37:00.341549 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:37:03 crc kubenswrapper[4696]: I0321 09:37:03.166696 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-xbjnb_f6174e0b-3b88-4452-ab2f-89905f5d09a8/cert-manager-controller/0.log" Mar 21 09:37:03 crc kubenswrapper[4696]: I0321 09:37:03.762452 4696 scope.go:117] "RemoveContainer" containerID="3884babbc72e0c022df25b381e932e30daa732179936a1c41eec0c731632e3d1" Mar 21 09:37:03 crc kubenswrapper[4696]: I0321 09:37:03.845137 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-f85xz_5a3acbc7-10ae-4287-9d73-15b97d7f0d71/cert-manager-cainjector/0.log" Mar 21 09:37:03 crc kubenswrapper[4696]: I0321 09:37:03.903713 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-rp8tc_e0d6313d-e133-464b-b5ea-33699c8280c0/cert-manager-webhook/0.log" Mar 21 09:37:15 crc kubenswrapper[4696]: I0321 09:37:15.464414 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-k9rdf_81abfcbd-f8b8-48d5-99b2-3cb7cc2d9160/nmstate-console-plugin/0.log" Mar 21 09:37:15 crc kubenswrapper[4696]: I0321 09:37:15.661153 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-r8wnr_be6dcd14-1d77-4c67-b527-29e7f46efc3e/nmstate-handler/0.log" Mar 21 09:37:15 crc kubenswrapper[4696]: I0321 09:37:15.733007 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-q8tc5_078ef7e2-dd73-486e-9213-804f5dbd26b2/kube-rbac-proxy/0.log" Mar 21 09:37:15 crc kubenswrapper[4696]: I0321 09:37:15.770882 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-q8tc5_078ef7e2-dd73-486e-9213-804f5dbd26b2/nmstate-metrics/0.log" Mar 21 09:37:15 crc kubenswrapper[4696]: I0321 09:37:15.869040 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-zc76d_3e70f161-07ef-4ed3-90be-6930cfb666a4/nmstate-operator/0.log" Mar 21 09:37:15 crc kubenswrapper[4696]: I0321 09:37:15.970108 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-q8s95_a351e45b-b9be-4898-8e07-c63cdd497245/nmstate-webhook/0.log" Mar 21 09:37:27 crc kubenswrapper[4696]: I0321 09:37:27.948776 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5685547f78-5x2wx_75d1fdab-70ac-46df-8b60-3e2f75985686/kube-rbac-proxy/0.log" Mar 21 09:37:28 crc kubenswrapper[4696]: I0321 09:37:28.063313 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5685547f78-5x2wx_75d1fdab-70ac-46df-8b60-3e2f75985686/manager/0.log" Mar 21 09:37:30 crc kubenswrapper[4696]: I0321 09:37:30.341867 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:37:30 crc kubenswrapper[4696]: I0321 09:37:30.342439 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:37:41 crc kubenswrapper[4696]: I0321 09:37:41.194696 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-8ff7d675-rs27t_527ae575-3dfd-4f0a-9af9-63668af676d3/prometheus-operator/0.log" Mar 21 09:37:41 crc kubenswrapper[4696]: I0321 09:37:41.399577 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-db48b98cb-78qsn_474e7a66-fb1d-448c-8297-b511ec7b6e7c/prometheus-operator-admission-webhook/0.log" Mar 21 09:37:41 crc kubenswrapper[4696]: I0321 09:37:41.403954 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-db48b98cb-l8q89_d99a53b9-d503-4410-9c7c-807fca5276bc/prometheus-operator-admission-webhook/0.log" Mar 21 09:37:41 crc kubenswrapper[4696]: I0321 09:37:41.606332 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-64bcb4cf6f-jjk5b_127ee197-aa20-41d3-ac26-e8aead566b64/perses-operator/0.log" Mar 21 09:37:41 crc kubenswrapper[4696]: I0321 09:37:41.625134 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-6dd7dd855f-sp4gg_e864a9cc-e270-4238-841b-77d0abad085d/operator/0.log" Mar 21 09:37:55 crc kubenswrapper[4696]: I0321 09:37:55.764745 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-lzh5g_3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c/kube-rbac-proxy/0.log" Mar 21 09:37:55 crc kubenswrapper[4696]: I0321 09:37:55.857640 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-lzh5g_3a676a7c-ba6e-4b0c-bfdd-fbdfabdcf30c/controller/0.log" Mar 21 09:37:56 crc kubenswrapper[4696]: I0321 09:37:56.003808 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-frr-files/0.log" Mar 21 09:37:56 crc kubenswrapper[4696]: I0321 09:37:56.244589 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-reloader/0.log" Mar 21 09:37:56 crc kubenswrapper[4696]: I0321 09:37:56.257338 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-reloader/0.log" Mar 21 09:37:56 crc kubenswrapper[4696]: I0321 09:37:56.275112 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-frr-files/0.log" Mar 21 09:37:56 crc kubenswrapper[4696]: I0321 09:37:56.284619 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-metrics/0.log" Mar 21 09:37:56 crc kubenswrapper[4696]: I0321 09:37:56.439694 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-frr-files/0.log" Mar 21 09:37:56 crc kubenswrapper[4696]: I0321 09:37:56.481059 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-metrics/0.log" Mar 21 09:37:56 crc kubenswrapper[4696]: I0321 09:37:56.498779 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-reloader/0.log" Mar 21 09:37:56 crc kubenswrapper[4696]: I0321 09:37:56.544752 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-metrics/0.log" Mar 21 09:37:56 crc kubenswrapper[4696]: I0321 09:37:56.684992 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-reloader/0.log" Mar 21 09:37:56 crc kubenswrapper[4696]: I0321 09:37:56.722583 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-metrics/0.log" Mar 21 09:37:56 crc kubenswrapper[4696]: I0321 09:37:56.769493 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/cp-frr-files/0.log" Mar 21 09:37:56 crc kubenswrapper[4696]: I0321 09:37:56.769832 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/controller/0.log" Mar 21 09:37:56 crc kubenswrapper[4696]: I0321 09:37:56.896049 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/frr-metrics/0.log" Mar 21 09:37:56 crc kubenswrapper[4696]: I0321 09:37:56.996496 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/kube-rbac-proxy/0.log" Mar 21 09:37:57 crc kubenswrapper[4696]: I0321 09:37:57.043429 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/kube-rbac-proxy-frr/0.log" Mar 21 09:37:57 crc kubenswrapper[4696]: I0321 09:37:57.209510 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/reloader/0.log" Mar 21 09:37:57 crc kubenswrapper[4696]: I0321 09:37:57.258218 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-gk8sz_69b8f173-b1f8-45e3-a6ae-ffb987fc709e/frr-k8s-webhook-server/0.log" Mar 21 09:37:57 crc kubenswrapper[4696]: I0321 09:37:57.674916 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-78876797bd-qsnl8_5fe21876-0bc8-489f-b862-6d48be898f5b/manager/0.log" Mar 21 09:37:57 crc kubenswrapper[4696]: I0321 09:37:57.838460 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7749f78475-cjlsc_dd16e737-b1c3-451b-a87c-90ec18d32dd3/webhook-server/0.log" Mar 21 09:37:57 crc kubenswrapper[4696]: I0321 09:37:57.880653 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zfggl_d8c2e94f-7eff-41dd-8e91-d6301ac9b84a/kube-rbac-proxy/0.log" Mar 21 09:37:58 crc kubenswrapper[4696]: I0321 09:37:58.499080 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zfggl_d8c2e94f-7eff-41dd-8e91-d6301ac9b84a/speaker/0.log" Mar 21 09:37:58 crc kubenswrapper[4696]: I0321 09:37:58.684004 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7v4k6_4f243b13-1046-485b-b726-469b5a2b283f/frr/0.log" Mar 21 09:38:00 crc kubenswrapper[4696]: I0321 09:38:00.161248 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568098-zct9z"] Mar 21 09:38:00 crc kubenswrapper[4696]: E0321 09:38:00.161731 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38f08644-f314-41b3-ab47-0f32a8de114c" containerName="registry-server" Mar 21 09:38:00 crc kubenswrapper[4696]: I0321 09:38:00.161748 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="38f08644-f314-41b3-ab47-0f32a8de114c" containerName="registry-server" Mar 21 09:38:00 crc kubenswrapper[4696]: E0321 09:38:00.161790 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38f08644-f314-41b3-ab47-0f32a8de114c" containerName="extract-content" Mar 21 09:38:00 crc kubenswrapper[4696]: I0321 09:38:00.161798 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="38f08644-f314-41b3-ab47-0f32a8de114c" containerName="extract-content" Mar 21 09:38:00 crc kubenswrapper[4696]: E0321 09:38:00.161862 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38f08644-f314-41b3-ab47-0f32a8de114c" containerName="extract-utilities" Mar 21 09:38:00 crc kubenswrapper[4696]: I0321 09:38:00.161870 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="38f08644-f314-41b3-ab47-0f32a8de114c" containerName="extract-utilities" Mar 21 09:38:00 crc kubenswrapper[4696]: I0321 09:38:00.162075 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="38f08644-f314-41b3-ab47-0f32a8de114c" containerName="registry-server" Mar 21 09:38:00 crc kubenswrapper[4696]: I0321 09:38:00.162969 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568098-zct9z" Mar 21 09:38:00 crc kubenswrapper[4696]: I0321 09:38:00.165909 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:38:00 crc kubenswrapper[4696]: I0321 09:38:00.166067 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:38:00 crc kubenswrapper[4696]: I0321 09:38:00.170748 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:38:00 crc kubenswrapper[4696]: I0321 09:38:00.186168 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568098-zct9z"] Mar 21 09:38:00 crc kubenswrapper[4696]: I0321 09:38:00.275693 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnbnv\" (UniqueName: \"kubernetes.io/projected/6f3b9dba-3a21-4101-92c3-59b62496539a-kube-api-access-hnbnv\") pod \"auto-csr-approver-29568098-zct9z\" (UID: \"6f3b9dba-3a21-4101-92c3-59b62496539a\") " pod="openshift-infra/auto-csr-approver-29568098-zct9z" Mar 21 09:38:00 crc kubenswrapper[4696]: I0321 09:38:00.341516 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:38:00 crc kubenswrapper[4696]: I0321 09:38:00.341583 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:38:00 crc kubenswrapper[4696]: I0321 09:38:00.341650 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 09:38:00 crc kubenswrapper[4696]: I0321 09:38:00.342584 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1967ae7e219ab85d3941d84ab56ee52381e3cd6376788e863ea1aa5183ce99b5"} pod="openshift-machine-config-operator/machine-config-daemon-z7srw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 09:38:00 crc kubenswrapper[4696]: I0321 09:38:00.342654 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" containerID="cri-o://1967ae7e219ab85d3941d84ab56ee52381e3cd6376788e863ea1aa5183ce99b5" gracePeriod=600 Mar 21 09:38:00 crc kubenswrapper[4696]: I0321 09:38:00.377985 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnbnv\" (UniqueName: \"kubernetes.io/projected/6f3b9dba-3a21-4101-92c3-59b62496539a-kube-api-access-hnbnv\") pod \"auto-csr-approver-29568098-zct9z\" (UID: \"6f3b9dba-3a21-4101-92c3-59b62496539a\") " pod="openshift-infra/auto-csr-approver-29568098-zct9z" Mar 21 09:38:00 crc kubenswrapper[4696]: I0321 09:38:00.400601 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnbnv\" (UniqueName: \"kubernetes.io/projected/6f3b9dba-3a21-4101-92c3-59b62496539a-kube-api-access-hnbnv\") pod \"auto-csr-approver-29568098-zct9z\" (UID: \"6f3b9dba-3a21-4101-92c3-59b62496539a\") " pod="openshift-infra/auto-csr-approver-29568098-zct9z" Mar 21 09:38:00 crc kubenswrapper[4696]: I0321 09:38:00.485482 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568098-zct9z" Mar 21 09:38:00 crc kubenswrapper[4696]: I0321 09:38:00.941665 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568098-zct9z"] Mar 21 09:38:00 crc kubenswrapper[4696]: W0321 09:38:00.946399 4696 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f3b9dba_3a21_4101_92c3_59b62496539a.slice/crio-dc10891a34265a6f865037da0d5d59f9f3aeccf33dc533644933666dba7683b6 WatchSource:0}: Error finding container dc10891a34265a6f865037da0d5d59f9f3aeccf33dc533644933666dba7683b6: Status 404 returned error can't find the container with id dc10891a34265a6f865037da0d5d59f9f3aeccf33dc533644933666dba7683b6 Mar 21 09:38:01 crc kubenswrapper[4696]: I0321 09:38:01.042430 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568098-zct9z" event={"ID":"6f3b9dba-3a21-4101-92c3-59b62496539a","Type":"ContainerStarted","Data":"dc10891a34265a6f865037da0d5d59f9f3aeccf33dc533644933666dba7683b6"} Mar 21 09:38:01 crc kubenswrapper[4696]: I0321 09:38:01.045512 4696 generic.go:334] "Generic (PLEG): container finished" podID="daaf227a-2305-495e-8495-a280abcd8e10" containerID="1967ae7e219ab85d3941d84ab56ee52381e3cd6376788e863ea1aa5183ce99b5" exitCode=0 Mar 21 09:38:01 crc kubenswrapper[4696]: I0321 09:38:01.045542 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerDied","Data":"1967ae7e219ab85d3941d84ab56ee52381e3cd6376788e863ea1aa5183ce99b5"} Mar 21 09:38:01 crc kubenswrapper[4696]: I0321 09:38:01.045562 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerStarted","Data":"30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5"} Mar 21 09:38:01 crc kubenswrapper[4696]: I0321 09:38:01.045579 4696 scope.go:117] "RemoveContainer" containerID="32d00a2a4e16c758505fb2e662996390a9563c70e183797967649432f285601b" Mar 21 09:38:03 crc kubenswrapper[4696]: I0321 09:38:03.065698 4696 generic.go:334] "Generic (PLEG): container finished" podID="6f3b9dba-3a21-4101-92c3-59b62496539a" containerID="a8834f1ba6ed59616dff8764718930d83b5348fcf1337e28a638c99401e5cb65" exitCode=0 Mar 21 09:38:03 crc kubenswrapper[4696]: I0321 09:38:03.065752 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568098-zct9z" event={"ID":"6f3b9dba-3a21-4101-92c3-59b62496539a","Type":"ContainerDied","Data":"a8834f1ba6ed59616dff8764718930d83b5348fcf1337e28a638c99401e5cb65"} Mar 21 09:38:04 crc kubenswrapper[4696]: I0321 09:38:04.512505 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568098-zct9z" Mar 21 09:38:04 crc kubenswrapper[4696]: I0321 09:38:04.675490 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnbnv\" (UniqueName: \"kubernetes.io/projected/6f3b9dba-3a21-4101-92c3-59b62496539a-kube-api-access-hnbnv\") pod \"6f3b9dba-3a21-4101-92c3-59b62496539a\" (UID: \"6f3b9dba-3a21-4101-92c3-59b62496539a\") " Mar 21 09:38:04 crc kubenswrapper[4696]: I0321 09:38:04.680805 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f3b9dba-3a21-4101-92c3-59b62496539a-kube-api-access-hnbnv" (OuterVolumeSpecName: "kube-api-access-hnbnv") pod "6f3b9dba-3a21-4101-92c3-59b62496539a" (UID: "6f3b9dba-3a21-4101-92c3-59b62496539a"). InnerVolumeSpecName "kube-api-access-hnbnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:38:04 crc kubenswrapper[4696]: I0321 09:38:04.779099 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnbnv\" (UniqueName: \"kubernetes.io/projected/6f3b9dba-3a21-4101-92c3-59b62496539a-kube-api-access-hnbnv\") on node \"crc\" DevicePath \"\"" Mar 21 09:38:05 crc kubenswrapper[4696]: I0321 09:38:05.087054 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568098-zct9z" event={"ID":"6f3b9dba-3a21-4101-92c3-59b62496539a","Type":"ContainerDied","Data":"dc10891a34265a6f865037da0d5d59f9f3aeccf33dc533644933666dba7683b6"} Mar 21 09:38:05 crc kubenswrapper[4696]: I0321 09:38:05.087334 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc10891a34265a6f865037da0d5d59f9f3aeccf33dc533644933666dba7683b6" Mar 21 09:38:05 crc kubenswrapper[4696]: I0321 09:38:05.087180 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568098-zct9z" Mar 21 09:38:05 crc kubenswrapper[4696]: I0321 09:38:05.584689 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568092-brhlh"] Mar 21 09:38:05 crc kubenswrapper[4696]: I0321 09:38:05.599398 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568092-brhlh"] Mar 21 09:38:06 crc kubenswrapper[4696]: I0321 09:38:06.547063 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f4afa53-dca8-4e98-9067-841e281f49c9" path="/var/lib/kubelet/pods/9f4afa53-dca8-4e98-9067-841e281f49c9/volumes" Mar 21 09:38:12 crc kubenswrapper[4696]: I0321 09:38:12.030515 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8_f46ecbe7-f1ca-4d24-975f-487b15fb2707/util/0.log" Mar 21 09:38:12 crc kubenswrapper[4696]: I0321 09:38:12.170973 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8_f46ecbe7-f1ca-4d24-975f-487b15fb2707/util/0.log" Mar 21 09:38:12 crc kubenswrapper[4696]: I0321 09:38:12.197041 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8_f46ecbe7-f1ca-4d24-975f-487b15fb2707/pull/0.log" Mar 21 09:38:12 crc kubenswrapper[4696]: I0321 09:38:12.281615 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8_f46ecbe7-f1ca-4d24-975f-487b15fb2707/pull/0.log" Mar 21 09:38:12 crc kubenswrapper[4696]: I0321 09:38:12.391942 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8_f46ecbe7-f1ca-4d24-975f-487b15fb2707/util/0.log" Mar 21 09:38:12 crc kubenswrapper[4696]: I0321 09:38:12.472313 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8_f46ecbe7-f1ca-4d24-975f-487b15fb2707/pull/0.log" Mar 21 09:38:12 crc kubenswrapper[4696]: I0321 09:38:12.492945 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874wltd8_f46ecbe7-f1ca-4d24-975f-487b15fb2707/extract/0.log" Mar 21 09:38:12 crc kubenswrapper[4696]: I0321 09:38:12.593026 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9_6ed1f144-c0f4-470e-9bc2-fc4ea36bc158/util/0.log" Mar 21 09:38:12 crc kubenswrapper[4696]: I0321 09:38:12.769728 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9_6ed1f144-c0f4-470e-9bc2-fc4ea36bc158/util/0.log" Mar 21 09:38:12 crc kubenswrapper[4696]: I0321 09:38:12.777239 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9_6ed1f144-c0f4-470e-9bc2-fc4ea36bc158/pull/0.log" Mar 21 09:38:12 crc kubenswrapper[4696]: I0321 09:38:12.790285 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9_6ed1f144-c0f4-470e-9bc2-fc4ea36bc158/pull/0.log" Mar 21 09:38:13 crc kubenswrapper[4696]: I0321 09:38:13.022143 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9_6ed1f144-c0f4-470e-9bc2-fc4ea36bc158/util/0.log" Mar 21 09:38:13 crc kubenswrapper[4696]: I0321 09:38:13.022206 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9_6ed1f144-c0f4-470e-9bc2-fc4ea36bc158/pull/0.log" Mar 21 09:38:13 crc kubenswrapper[4696]: I0321 09:38:13.023075 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1f52l9_6ed1f144-c0f4-470e-9bc2-fc4ea36bc158/extract/0.log" Mar 21 09:38:13 crc kubenswrapper[4696]: I0321 09:38:13.218453 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5_e4b15b92-b219-4ad6-a9ab-7d05a7130c92/util/0.log" Mar 21 09:38:13 crc kubenswrapper[4696]: I0321 09:38:13.433611 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5_e4b15b92-b219-4ad6-a9ab-7d05a7130c92/pull/0.log" Mar 21 09:38:13 crc kubenswrapper[4696]: I0321 09:38:13.436707 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5_e4b15b92-b219-4ad6-a9ab-7d05a7130c92/util/0.log" Mar 21 09:38:13 crc kubenswrapper[4696]: I0321 09:38:13.507520 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5_e4b15b92-b219-4ad6-a9ab-7d05a7130c92/pull/0.log" Mar 21 09:38:13 crc kubenswrapper[4696]: I0321 09:38:13.626621 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5_e4b15b92-b219-4ad6-a9ab-7d05a7130c92/util/0.log" Mar 21 09:38:13 crc kubenswrapper[4696]: I0321 09:38:13.651314 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5_e4b15b92-b219-4ad6-a9ab-7d05a7130c92/pull/0.log" Mar 21 09:38:13 crc kubenswrapper[4696]: I0321 09:38:13.726155 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726s7vw5_e4b15b92-b219-4ad6-a9ab-7d05a7130c92/extract/0.log" Mar 21 09:38:13 crc kubenswrapper[4696]: I0321 09:38:13.892307 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh_8c8ec8ba-d404-4dca-aac7-3619ef7dbd75/util/0.log" Mar 21 09:38:14 crc kubenswrapper[4696]: I0321 09:38:14.004874 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh_8c8ec8ba-d404-4dca-aac7-3619ef7dbd75/util/0.log" Mar 21 09:38:14 crc kubenswrapper[4696]: I0321 09:38:14.049910 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh_8c8ec8ba-d404-4dca-aac7-3619ef7dbd75/pull/0.log" Mar 21 09:38:14 crc kubenswrapper[4696]: I0321 09:38:14.066829 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh_8c8ec8ba-d404-4dca-aac7-3619ef7dbd75/pull/0.log" Mar 21 09:38:14 crc kubenswrapper[4696]: I0321 09:38:14.316551 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh_8c8ec8ba-d404-4dca-aac7-3619ef7dbd75/util/0.log" Mar 21 09:38:14 crc kubenswrapper[4696]: I0321 09:38:14.440775 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh_8c8ec8ba-d404-4dca-aac7-3619ef7dbd75/extract/0.log" Mar 21 09:38:14 crc kubenswrapper[4696]: I0321 09:38:14.485122 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_b42be0e84d359797ca2a6e73d6fbaa3b214e20650a24a735d2f27f15fc4bxzh_8c8ec8ba-d404-4dca-aac7-3619ef7dbd75/pull/0.log" Mar 21 09:38:14 crc kubenswrapper[4696]: I0321 09:38:14.573853 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5vlg_e175ab46-d3f4-4051-a86e-dcf2727b9efd/extract-utilities/0.log" Mar 21 09:38:14 crc kubenswrapper[4696]: I0321 09:38:14.759116 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5vlg_e175ab46-d3f4-4051-a86e-dcf2727b9efd/extract-content/0.log" Mar 21 09:38:14 crc kubenswrapper[4696]: I0321 09:38:14.764019 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5vlg_e175ab46-d3f4-4051-a86e-dcf2727b9efd/extract-content/0.log" Mar 21 09:38:14 crc kubenswrapper[4696]: I0321 09:38:14.793082 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5vlg_e175ab46-d3f4-4051-a86e-dcf2727b9efd/extract-utilities/0.log" Mar 21 09:38:14 crc kubenswrapper[4696]: I0321 09:38:14.964530 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5vlg_e175ab46-d3f4-4051-a86e-dcf2727b9efd/extract-utilities/0.log" Mar 21 09:38:15 crc kubenswrapper[4696]: I0321 09:38:15.054808 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5vlg_e175ab46-d3f4-4051-a86e-dcf2727b9efd/extract-content/0.log" Mar 21 09:38:15 crc kubenswrapper[4696]: I0321 09:38:15.230355 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rwzg5_1050b6cc-b45d-42cb-8d54-7104988c8225/extract-utilities/0.log" Mar 21 09:38:15 crc kubenswrapper[4696]: I0321 09:38:15.457618 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rwzg5_1050b6cc-b45d-42cb-8d54-7104988c8225/extract-utilities/0.log" Mar 21 09:38:15 crc kubenswrapper[4696]: I0321 09:38:15.467067 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rwzg5_1050b6cc-b45d-42cb-8d54-7104988c8225/extract-content/0.log" Mar 21 09:38:15 crc kubenswrapper[4696]: I0321 09:38:15.507280 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rwzg5_1050b6cc-b45d-42cb-8d54-7104988c8225/extract-content/0.log" Mar 21 09:38:15 crc kubenswrapper[4696]: I0321 09:38:15.620207 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5vlg_e175ab46-d3f4-4051-a86e-dcf2727b9efd/registry-server/0.log" Mar 21 09:38:15 crc kubenswrapper[4696]: I0321 09:38:15.722540 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rwzg5_1050b6cc-b45d-42cb-8d54-7104988c8225/extract-utilities/0.log" Mar 21 09:38:15 crc kubenswrapper[4696]: I0321 09:38:15.726746 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rwzg5_1050b6cc-b45d-42cb-8d54-7104988c8225/extract-content/0.log" Mar 21 09:38:16 crc kubenswrapper[4696]: I0321 09:38:16.056194 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-w6xbt_d45283e8-56c7-4a2e-8a7e-88e459bb1c04/marketplace-operator/0.log" Mar 21 09:38:16 crc kubenswrapper[4696]: I0321 09:38:16.200854 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mmkz6_5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77/extract-utilities/0.log" Mar 21 09:38:16 crc kubenswrapper[4696]: I0321 09:38:16.392393 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rwzg5_1050b6cc-b45d-42cb-8d54-7104988c8225/registry-server/0.log" Mar 21 09:38:16 crc kubenswrapper[4696]: I0321 09:38:16.761564 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mmkz6_5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77/extract-content/0.log" Mar 21 09:38:16 crc kubenswrapper[4696]: I0321 09:38:16.765386 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mmkz6_5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77/extract-content/0.log" Mar 21 09:38:16 crc kubenswrapper[4696]: I0321 09:38:16.780435 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mmkz6_5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77/extract-utilities/0.log" Mar 21 09:38:16 crc kubenswrapper[4696]: I0321 09:38:16.920209 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mmkz6_5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77/extract-content/0.log" Mar 21 09:38:16 crc kubenswrapper[4696]: I0321 09:38:16.926888 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mmkz6_5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77/extract-utilities/0.log" Mar 21 09:38:16 crc kubenswrapper[4696]: I0321 09:38:16.995383 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6f4n_b49f2cd9-95cf-4749-889e-c47d027c253b/extract-utilities/0.log" Mar 21 09:38:17 crc kubenswrapper[4696]: I0321 09:38:17.079542 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mmkz6_5cb0e56a-eac9-4103-8d0c-e5ebc1c5dd77/registry-server/0.log" Mar 21 09:38:17 crc kubenswrapper[4696]: I0321 09:38:17.160728 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6f4n_b49f2cd9-95cf-4749-889e-c47d027c253b/extract-utilities/0.log" Mar 21 09:38:17 crc kubenswrapper[4696]: I0321 09:38:17.194794 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6f4n_b49f2cd9-95cf-4749-889e-c47d027c253b/extract-content/0.log" Mar 21 09:38:17 crc kubenswrapper[4696]: I0321 09:38:17.195884 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6f4n_b49f2cd9-95cf-4749-889e-c47d027c253b/extract-content/0.log" Mar 21 09:38:17 crc kubenswrapper[4696]: I0321 09:38:17.366370 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6f4n_b49f2cd9-95cf-4749-889e-c47d027c253b/extract-content/0.log" Mar 21 09:38:17 crc kubenswrapper[4696]: I0321 09:38:17.372972 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6f4n_b49f2cd9-95cf-4749-889e-c47d027c253b/extract-utilities/0.log" Mar 21 09:38:18 crc kubenswrapper[4696]: I0321 09:38:18.103619 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6f4n_b49f2cd9-95cf-4749-889e-c47d027c253b/registry-server/0.log" Mar 21 09:38:30 crc kubenswrapper[4696]: I0321 09:38:30.832191 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-8ff7d675-rs27t_527ae575-3dfd-4f0a-9af9-63668af676d3/prometheus-operator/0.log" Mar 21 09:38:30 crc kubenswrapper[4696]: I0321 09:38:30.895342 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-db48b98cb-78qsn_474e7a66-fb1d-448c-8297-b511ec7b6e7c/prometheus-operator-admission-webhook/0.log" Mar 21 09:38:30 crc kubenswrapper[4696]: I0321 09:38:30.916065 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-db48b98cb-l8q89_d99a53b9-d503-4410-9c7c-807fca5276bc/prometheus-operator-admission-webhook/0.log" Mar 21 09:38:31 crc kubenswrapper[4696]: I0321 09:38:31.072379 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-6dd7dd855f-sp4gg_e864a9cc-e270-4238-841b-77d0abad085d/operator/0.log" Mar 21 09:38:31 crc kubenswrapper[4696]: I0321 09:38:31.119299 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-64bcb4cf6f-jjk5b_127ee197-aa20-41d3-ac26-e8aead566b64/perses-operator/0.log" Mar 21 09:38:45 crc kubenswrapper[4696]: I0321 09:38:45.827683 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5685547f78-5x2wx_75d1fdab-70ac-46df-8b60-3e2f75985686/kube-rbac-proxy/0.log" Mar 21 09:38:45 crc kubenswrapper[4696]: I0321 09:38:45.841948 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5685547f78-5x2wx_75d1fdab-70ac-46df-8b60-3e2f75985686/manager/0.log" Mar 21 09:39:03 crc kubenswrapper[4696]: I0321 09:39:03.908832 4696 scope.go:117] "RemoveContainer" containerID="441f4831eadfbe70188580b2a40c0f5fe7dcdc96e70587660084294bcaa63505" Mar 21 09:40:00 crc kubenswrapper[4696]: I0321 09:40:00.153477 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568100-82jfn"] Mar 21 09:40:00 crc kubenswrapper[4696]: E0321 09:40:00.155152 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f3b9dba-3a21-4101-92c3-59b62496539a" containerName="oc" Mar 21 09:40:00 crc kubenswrapper[4696]: I0321 09:40:00.155181 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f3b9dba-3a21-4101-92c3-59b62496539a" containerName="oc" Mar 21 09:40:00 crc kubenswrapper[4696]: I0321 09:40:00.155557 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f3b9dba-3a21-4101-92c3-59b62496539a" containerName="oc" Mar 21 09:40:00 crc kubenswrapper[4696]: I0321 09:40:00.156900 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568100-82jfn" Mar 21 09:40:00 crc kubenswrapper[4696]: I0321 09:40:00.160288 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:40:00 crc kubenswrapper[4696]: I0321 09:40:00.160306 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:40:00 crc kubenswrapper[4696]: I0321 09:40:00.161266 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:40:00 crc kubenswrapper[4696]: I0321 09:40:00.169936 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568100-82jfn"] Mar 21 09:40:00 crc kubenswrapper[4696]: I0321 09:40:00.286315 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccvvk\" (UniqueName: \"kubernetes.io/projected/a6dd8c3f-a5d2-44c3-b196-90631d82cdc3-kube-api-access-ccvvk\") pod \"auto-csr-approver-29568100-82jfn\" (UID: \"a6dd8c3f-a5d2-44c3-b196-90631d82cdc3\") " pod="openshift-infra/auto-csr-approver-29568100-82jfn" Mar 21 09:40:00 crc kubenswrapper[4696]: I0321 09:40:00.341347 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:40:00 crc kubenswrapper[4696]: I0321 09:40:00.341788 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:40:00 crc kubenswrapper[4696]: I0321 09:40:00.388642 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccvvk\" (UniqueName: \"kubernetes.io/projected/a6dd8c3f-a5d2-44c3-b196-90631d82cdc3-kube-api-access-ccvvk\") pod \"auto-csr-approver-29568100-82jfn\" (UID: \"a6dd8c3f-a5d2-44c3-b196-90631d82cdc3\") " pod="openshift-infra/auto-csr-approver-29568100-82jfn" Mar 21 09:40:00 crc kubenswrapper[4696]: I0321 09:40:00.412439 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccvvk\" (UniqueName: \"kubernetes.io/projected/a6dd8c3f-a5d2-44c3-b196-90631d82cdc3-kube-api-access-ccvvk\") pod \"auto-csr-approver-29568100-82jfn\" (UID: \"a6dd8c3f-a5d2-44c3-b196-90631d82cdc3\") " pod="openshift-infra/auto-csr-approver-29568100-82jfn" Mar 21 09:40:00 crc kubenswrapper[4696]: I0321 09:40:00.480517 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568100-82jfn" Mar 21 09:40:00 crc kubenswrapper[4696]: I0321 09:40:00.984652 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568100-82jfn"] Mar 21 09:40:01 crc kubenswrapper[4696]: I0321 09:40:01.255536 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568100-82jfn" event={"ID":"a6dd8c3f-a5d2-44c3-b196-90631d82cdc3","Type":"ContainerStarted","Data":"0223b4fc2eb99c631ea355ff9dcb3be106b49b026efe8ec08d5f9dcfbd3ccd0f"} Mar 21 09:40:02 crc kubenswrapper[4696]: I0321 09:40:02.265773 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568100-82jfn" event={"ID":"a6dd8c3f-a5d2-44c3-b196-90631d82cdc3","Type":"ContainerStarted","Data":"6dfd97b916e7d1383bc1cbc7f92bfdfe4f139c0bc6b4228267414eba46d3092c"} Mar 21 09:40:02 crc kubenswrapper[4696]: I0321 09:40:02.295068 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29568100-82jfn" podStartSLOduration=1.331213915 podStartE2EDuration="2.295045657s" podCreationTimestamp="2026-03-21 09:40:00 +0000 UTC" firstStartedPulling="2026-03-21 09:40:00.983266487 +0000 UTC m=+4335.104147200" lastFinishedPulling="2026-03-21 09:40:01.947098229 +0000 UTC m=+4336.067978942" observedRunningTime="2026-03-21 09:40:02.283444496 +0000 UTC m=+4336.404325209" watchObservedRunningTime="2026-03-21 09:40:02.295045657 +0000 UTC m=+4336.415926380" Mar 21 09:40:03 crc kubenswrapper[4696]: I0321 09:40:03.276267 4696 generic.go:334] "Generic (PLEG): container finished" podID="a6dd8c3f-a5d2-44c3-b196-90631d82cdc3" containerID="6dfd97b916e7d1383bc1cbc7f92bfdfe4f139c0bc6b4228267414eba46d3092c" exitCode=0 Mar 21 09:40:03 crc kubenswrapper[4696]: I0321 09:40:03.276364 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568100-82jfn" event={"ID":"a6dd8c3f-a5d2-44c3-b196-90631d82cdc3","Type":"ContainerDied","Data":"6dfd97b916e7d1383bc1cbc7f92bfdfe4f139c0bc6b4228267414eba46d3092c"} Mar 21 09:40:04 crc kubenswrapper[4696]: I0321 09:40:04.010281 4696 scope.go:117] "RemoveContainer" containerID="2de4c3370996ba3ca8ec495d1be764876de2259f76adaca7b54ed497311aabbb" Mar 21 09:40:04 crc kubenswrapper[4696]: I0321 09:40:04.642184 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568100-82jfn" Mar 21 09:40:04 crc kubenswrapper[4696]: I0321 09:40:04.801722 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccvvk\" (UniqueName: \"kubernetes.io/projected/a6dd8c3f-a5d2-44c3-b196-90631d82cdc3-kube-api-access-ccvvk\") pod \"a6dd8c3f-a5d2-44c3-b196-90631d82cdc3\" (UID: \"a6dd8c3f-a5d2-44c3-b196-90631d82cdc3\") " Mar 21 09:40:04 crc kubenswrapper[4696]: I0321 09:40:04.820909 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6dd8c3f-a5d2-44c3-b196-90631d82cdc3-kube-api-access-ccvvk" (OuterVolumeSpecName: "kube-api-access-ccvvk") pod "a6dd8c3f-a5d2-44c3-b196-90631d82cdc3" (UID: "a6dd8c3f-a5d2-44c3-b196-90631d82cdc3"). InnerVolumeSpecName "kube-api-access-ccvvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:40:04 crc kubenswrapper[4696]: I0321 09:40:04.903965 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccvvk\" (UniqueName: \"kubernetes.io/projected/a6dd8c3f-a5d2-44c3-b196-90631d82cdc3-kube-api-access-ccvvk\") on node \"crc\" DevicePath \"\"" Mar 21 09:40:05 crc kubenswrapper[4696]: I0321 09:40:05.296222 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568100-82jfn" event={"ID":"a6dd8c3f-a5d2-44c3-b196-90631d82cdc3","Type":"ContainerDied","Data":"0223b4fc2eb99c631ea355ff9dcb3be106b49b026efe8ec08d5f9dcfbd3ccd0f"} Mar 21 09:40:05 crc kubenswrapper[4696]: I0321 09:40:05.296267 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0223b4fc2eb99c631ea355ff9dcb3be106b49b026efe8ec08d5f9dcfbd3ccd0f" Mar 21 09:40:05 crc kubenswrapper[4696]: I0321 09:40:05.296317 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568100-82jfn" Mar 21 09:40:05 crc kubenswrapper[4696]: I0321 09:40:05.353569 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568094-92m6p"] Mar 21 09:40:05 crc kubenswrapper[4696]: I0321 09:40:05.367169 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568094-92m6p"] Mar 21 09:40:06 crc kubenswrapper[4696]: I0321 09:40:06.549452 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0da3d36c-c518-48eb-8ea5-df44b5681206" path="/var/lib/kubelet/pods/0da3d36c-c518-48eb-8ea5-df44b5681206/volumes" Mar 21 09:40:30 crc kubenswrapper[4696]: I0321 09:40:30.341192 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:40:30 crc kubenswrapper[4696]: I0321 09:40:30.341638 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:40:39 crc kubenswrapper[4696]: I0321 09:40:39.639800 4696 generic.go:334] "Generic (PLEG): container finished" podID="f55d5321-1e8a-42bc-8c6e-17a2d032639b" containerID="33eae8dd16249faa0e0d07108fbe6627adf42580bb3a6c3cb04606153d5bd0b6" exitCode=0 Mar 21 09:40:39 crc kubenswrapper[4696]: I0321 09:40:39.639922 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gw6qx/must-gather-mn5rt" event={"ID":"f55d5321-1e8a-42bc-8c6e-17a2d032639b","Type":"ContainerDied","Data":"33eae8dd16249faa0e0d07108fbe6627adf42580bb3a6c3cb04606153d5bd0b6"} Mar 21 09:40:39 crc kubenswrapper[4696]: I0321 09:40:39.641299 4696 scope.go:117] "RemoveContainer" containerID="33eae8dd16249faa0e0d07108fbe6627adf42580bb3a6c3cb04606153d5bd0b6" Mar 21 09:40:39 crc kubenswrapper[4696]: I0321 09:40:39.745955 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gw6qx_must-gather-mn5rt_f55d5321-1e8a-42bc-8c6e-17a2d032639b/gather/0.log" Mar 21 09:40:52 crc kubenswrapper[4696]: I0321 09:40:52.370139 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gw6qx/must-gather-mn5rt"] Mar 21 09:40:52 crc kubenswrapper[4696]: I0321 09:40:52.370903 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-gw6qx/must-gather-mn5rt" podUID="f55d5321-1e8a-42bc-8c6e-17a2d032639b" containerName="copy" containerID="cri-o://731bf94972a15a9640f391e81eb4eb80c2816d07d220cf2947b312112d371411" gracePeriod=2 Mar 21 09:40:52 crc kubenswrapper[4696]: I0321 09:40:52.388397 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gw6qx/must-gather-mn5rt"] Mar 21 09:40:53 crc kubenswrapper[4696]: I0321 09:40:53.245220 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gw6qx_must-gather-mn5rt_f55d5321-1e8a-42bc-8c6e-17a2d032639b/copy/0.log" Mar 21 09:40:53 crc kubenswrapper[4696]: I0321 09:40:53.245882 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gw6qx/must-gather-mn5rt" Mar 21 09:40:53 crc kubenswrapper[4696]: I0321 09:40:53.412949 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6wlz\" (UniqueName: \"kubernetes.io/projected/f55d5321-1e8a-42bc-8c6e-17a2d032639b-kube-api-access-j6wlz\") pod \"f55d5321-1e8a-42bc-8c6e-17a2d032639b\" (UID: \"f55d5321-1e8a-42bc-8c6e-17a2d032639b\") " Mar 21 09:40:53 crc kubenswrapper[4696]: I0321 09:40:53.413061 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f55d5321-1e8a-42bc-8c6e-17a2d032639b-must-gather-output\") pod \"f55d5321-1e8a-42bc-8c6e-17a2d032639b\" (UID: \"f55d5321-1e8a-42bc-8c6e-17a2d032639b\") " Mar 21 09:40:53 crc kubenswrapper[4696]: I0321 09:40:53.468069 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f55d5321-1e8a-42bc-8c6e-17a2d032639b-kube-api-access-j6wlz" (OuterVolumeSpecName: "kube-api-access-j6wlz") pod "f55d5321-1e8a-42bc-8c6e-17a2d032639b" (UID: "f55d5321-1e8a-42bc-8c6e-17a2d032639b"). InnerVolumeSpecName "kube-api-access-j6wlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:40:53 crc kubenswrapper[4696]: I0321 09:40:53.515871 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6wlz\" (UniqueName: \"kubernetes.io/projected/f55d5321-1e8a-42bc-8c6e-17a2d032639b-kube-api-access-j6wlz\") on node \"crc\" DevicePath \"\"" Mar 21 09:40:53 crc kubenswrapper[4696]: I0321 09:40:53.599779 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f55d5321-1e8a-42bc-8c6e-17a2d032639b-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "f55d5321-1e8a-42bc-8c6e-17a2d032639b" (UID: "f55d5321-1e8a-42bc-8c6e-17a2d032639b"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:40:53 crc kubenswrapper[4696]: I0321 09:40:53.617761 4696 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f55d5321-1e8a-42bc-8c6e-17a2d032639b-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 21 09:40:53 crc kubenswrapper[4696]: I0321 09:40:53.776783 4696 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gw6qx_must-gather-mn5rt_f55d5321-1e8a-42bc-8c6e-17a2d032639b/copy/0.log" Mar 21 09:40:53 crc kubenswrapper[4696]: I0321 09:40:53.777580 4696 generic.go:334] "Generic (PLEG): container finished" podID="f55d5321-1e8a-42bc-8c6e-17a2d032639b" containerID="731bf94972a15a9640f391e81eb4eb80c2816d07d220cf2947b312112d371411" exitCode=143 Mar 21 09:40:53 crc kubenswrapper[4696]: I0321 09:40:53.777643 4696 scope.go:117] "RemoveContainer" containerID="731bf94972a15a9640f391e81eb4eb80c2816d07d220cf2947b312112d371411" Mar 21 09:40:53 crc kubenswrapper[4696]: I0321 09:40:53.777648 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gw6qx/must-gather-mn5rt" Mar 21 09:40:53 crc kubenswrapper[4696]: I0321 09:40:53.798514 4696 scope.go:117] "RemoveContainer" containerID="33eae8dd16249faa0e0d07108fbe6627adf42580bb3a6c3cb04606153d5bd0b6" Mar 21 09:40:53 crc kubenswrapper[4696]: I0321 09:40:53.835615 4696 scope.go:117] "RemoveContainer" containerID="731bf94972a15a9640f391e81eb4eb80c2816d07d220cf2947b312112d371411" Mar 21 09:40:53 crc kubenswrapper[4696]: E0321 09:40:53.836241 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"731bf94972a15a9640f391e81eb4eb80c2816d07d220cf2947b312112d371411\": container with ID starting with 731bf94972a15a9640f391e81eb4eb80c2816d07d220cf2947b312112d371411 not found: ID does not exist" containerID="731bf94972a15a9640f391e81eb4eb80c2816d07d220cf2947b312112d371411" Mar 21 09:40:53 crc kubenswrapper[4696]: I0321 09:40:53.836272 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"731bf94972a15a9640f391e81eb4eb80c2816d07d220cf2947b312112d371411"} err="failed to get container status \"731bf94972a15a9640f391e81eb4eb80c2816d07d220cf2947b312112d371411\": rpc error: code = NotFound desc = could not find container \"731bf94972a15a9640f391e81eb4eb80c2816d07d220cf2947b312112d371411\": container with ID starting with 731bf94972a15a9640f391e81eb4eb80c2816d07d220cf2947b312112d371411 not found: ID does not exist" Mar 21 09:40:53 crc kubenswrapper[4696]: I0321 09:40:53.836293 4696 scope.go:117] "RemoveContainer" containerID="33eae8dd16249faa0e0d07108fbe6627adf42580bb3a6c3cb04606153d5bd0b6" Mar 21 09:40:53 crc kubenswrapper[4696]: E0321 09:40:53.836698 4696 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33eae8dd16249faa0e0d07108fbe6627adf42580bb3a6c3cb04606153d5bd0b6\": container with ID starting with 33eae8dd16249faa0e0d07108fbe6627adf42580bb3a6c3cb04606153d5bd0b6 not found: ID does not exist" containerID="33eae8dd16249faa0e0d07108fbe6627adf42580bb3a6c3cb04606153d5bd0b6" Mar 21 09:40:53 crc kubenswrapper[4696]: I0321 09:40:53.836751 4696 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33eae8dd16249faa0e0d07108fbe6627adf42580bb3a6c3cb04606153d5bd0b6"} err="failed to get container status \"33eae8dd16249faa0e0d07108fbe6627adf42580bb3a6c3cb04606153d5bd0b6\": rpc error: code = NotFound desc = could not find container \"33eae8dd16249faa0e0d07108fbe6627adf42580bb3a6c3cb04606153d5bd0b6\": container with ID starting with 33eae8dd16249faa0e0d07108fbe6627adf42580bb3a6c3cb04606153d5bd0b6 not found: ID does not exist" Mar 21 09:40:54 crc kubenswrapper[4696]: I0321 09:40:54.546682 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f55d5321-1e8a-42bc-8c6e-17a2d032639b" path="/var/lib/kubelet/pods/f55d5321-1e8a-42bc-8c6e-17a2d032639b/volumes" Mar 21 09:41:00 crc kubenswrapper[4696]: I0321 09:41:00.341622 4696 patch_prober.go:28] interesting pod/machine-config-daemon-z7srw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 09:41:00 crc kubenswrapper[4696]: I0321 09:41:00.341984 4696 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 09:41:00 crc kubenswrapper[4696]: I0321 09:41:00.342025 4696 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" Mar 21 09:41:00 crc kubenswrapper[4696]: I0321 09:41:00.342753 4696 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5"} pod="openshift-machine-config-operator/machine-config-daemon-z7srw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 09:41:00 crc kubenswrapper[4696]: I0321 09:41:00.342804 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" containerName="machine-config-daemon" containerID="cri-o://30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" gracePeriod=600 Mar 21 09:41:00 crc kubenswrapper[4696]: E0321 09:41:00.487493 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:41:00 crc kubenswrapper[4696]: I0321 09:41:00.860059 4696 generic.go:334] "Generic (PLEG): container finished" podID="daaf227a-2305-495e-8495-a280abcd8e10" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" exitCode=0 Mar 21 09:41:00 crc kubenswrapper[4696]: I0321 09:41:00.860106 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" event={"ID":"daaf227a-2305-495e-8495-a280abcd8e10","Type":"ContainerDied","Data":"30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5"} Mar 21 09:41:00 crc kubenswrapper[4696]: I0321 09:41:00.860142 4696 scope.go:117] "RemoveContainer" containerID="1967ae7e219ab85d3941d84ab56ee52381e3cd6376788e863ea1aa5183ce99b5" Mar 21 09:41:00 crc kubenswrapper[4696]: I0321 09:41:00.860921 4696 scope.go:117] "RemoveContainer" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" Mar 21 09:41:00 crc kubenswrapper[4696]: E0321 09:41:00.861321 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:41:04 crc kubenswrapper[4696]: I0321 09:41:04.073614 4696 scope.go:117] "RemoveContainer" containerID="50f64b41930601c15e20c2cfcb0ef014df74cb5f1d6808c3a37a76a9639e45db" Mar 21 09:41:04 crc kubenswrapper[4696]: I0321 09:41:04.100525 4696 scope.go:117] "RemoveContainer" containerID="8338ad4703d22b6bd59548fb354a04e025e7cd58efb4cef501fd91a8c0196c4b" Mar 21 09:41:04 crc kubenswrapper[4696]: I0321 09:41:04.150077 4696 scope.go:117] "RemoveContainer" containerID="46461fd2615b0d286720ddf99d31e4e1206699d9f8b5dc98fadfed16e5f9aec4" Mar 21 09:41:04 crc kubenswrapper[4696]: I0321 09:41:04.187125 4696 scope.go:117] "RemoveContainer" containerID="520e45f892d998ba05a89db45b717c566ab1e6c039b8a1a0902960b68a391256" Mar 21 09:41:04 crc kubenswrapper[4696]: I0321 09:41:04.226633 4696 scope.go:117] "RemoveContainer" containerID="9883203e663db62ec9ef4b42bfc76f4d97cbedb172227e93b8f0b89ffb0a72cd" Mar 21 09:41:14 crc kubenswrapper[4696]: I0321 09:41:14.535736 4696 scope.go:117] "RemoveContainer" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" Mar 21 09:41:14 crc kubenswrapper[4696]: E0321 09:41:14.539149 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:41:25 crc kubenswrapper[4696]: I0321 09:41:25.536158 4696 scope.go:117] "RemoveContainer" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" Mar 21 09:41:25 crc kubenswrapper[4696]: E0321 09:41:25.537576 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:41:36 crc kubenswrapper[4696]: I0321 09:41:36.543414 4696 scope.go:117] "RemoveContainer" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" Mar 21 09:41:36 crc kubenswrapper[4696]: E0321 09:41:36.546290 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:41:50 crc kubenswrapper[4696]: I0321 09:41:50.535106 4696 scope.go:117] "RemoveContainer" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" Mar 21 09:41:50 crc kubenswrapper[4696]: E0321 09:41:50.536167 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:42:00 crc kubenswrapper[4696]: I0321 09:42:00.145073 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568102-pjxb7"] Mar 21 09:42:00 crc kubenswrapper[4696]: E0321 09:42:00.146210 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6dd8c3f-a5d2-44c3-b196-90631d82cdc3" containerName="oc" Mar 21 09:42:00 crc kubenswrapper[4696]: I0321 09:42:00.146228 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6dd8c3f-a5d2-44c3-b196-90631d82cdc3" containerName="oc" Mar 21 09:42:00 crc kubenswrapper[4696]: E0321 09:42:00.146257 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f55d5321-1e8a-42bc-8c6e-17a2d032639b" containerName="gather" Mar 21 09:42:00 crc kubenswrapper[4696]: I0321 09:42:00.146266 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f55d5321-1e8a-42bc-8c6e-17a2d032639b" containerName="gather" Mar 21 09:42:00 crc kubenswrapper[4696]: E0321 09:42:00.146286 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f55d5321-1e8a-42bc-8c6e-17a2d032639b" containerName="copy" Mar 21 09:42:00 crc kubenswrapper[4696]: I0321 09:42:00.146294 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="f55d5321-1e8a-42bc-8c6e-17a2d032639b" containerName="copy" Mar 21 09:42:00 crc kubenswrapper[4696]: I0321 09:42:00.146586 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6dd8c3f-a5d2-44c3-b196-90631d82cdc3" containerName="oc" Mar 21 09:42:00 crc kubenswrapper[4696]: I0321 09:42:00.146605 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f55d5321-1e8a-42bc-8c6e-17a2d032639b" containerName="copy" Mar 21 09:42:00 crc kubenswrapper[4696]: I0321 09:42:00.146627 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="f55d5321-1e8a-42bc-8c6e-17a2d032639b" containerName="gather" Mar 21 09:42:00 crc kubenswrapper[4696]: I0321 09:42:00.147606 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568102-pjxb7" Mar 21 09:42:00 crc kubenswrapper[4696]: I0321 09:42:00.149988 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:42:00 crc kubenswrapper[4696]: I0321 09:42:00.150047 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:42:00 crc kubenswrapper[4696]: I0321 09:42:00.151181 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:42:00 crc kubenswrapper[4696]: I0321 09:42:00.156569 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568102-pjxb7"] Mar 21 09:42:00 crc kubenswrapper[4696]: I0321 09:42:00.207061 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnjk4\" (UniqueName: \"kubernetes.io/projected/5aa85936-31bf-4cdb-97fe-fa4e606c3c5f-kube-api-access-mnjk4\") pod \"auto-csr-approver-29568102-pjxb7\" (UID: \"5aa85936-31bf-4cdb-97fe-fa4e606c3c5f\") " pod="openshift-infra/auto-csr-approver-29568102-pjxb7" Mar 21 09:42:00 crc kubenswrapper[4696]: I0321 09:42:00.310053 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnjk4\" (UniqueName: \"kubernetes.io/projected/5aa85936-31bf-4cdb-97fe-fa4e606c3c5f-kube-api-access-mnjk4\") pod \"auto-csr-approver-29568102-pjxb7\" (UID: \"5aa85936-31bf-4cdb-97fe-fa4e606c3c5f\") " pod="openshift-infra/auto-csr-approver-29568102-pjxb7" Mar 21 09:42:00 crc kubenswrapper[4696]: I0321 09:42:00.326573 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnjk4\" (UniqueName: \"kubernetes.io/projected/5aa85936-31bf-4cdb-97fe-fa4e606c3c5f-kube-api-access-mnjk4\") pod \"auto-csr-approver-29568102-pjxb7\" (UID: \"5aa85936-31bf-4cdb-97fe-fa4e606c3c5f\") " pod="openshift-infra/auto-csr-approver-29568102-pjxb7" Mar 21 09:42:00 crc kubenswrapper[4696]: I0321 09:42:00.504680 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568102-pjxb7" Mar 21 09:42:00 crc kubenswrapper[4696]: I0321 09:42:00.964159 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568102-pjxb7"] Mar 21 09:42:00 crc kubenswrapper[4696]: I0321 09:42:00.965842 4696 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 09:42:01 crc kubenswrapper[4696]: I0321 09:42:01.415434 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568102-pjxb7" event={"ID":"5aa85936-31bf-4cdb-97fe-fa4e606c3c5f","Type":"ContainerStarted","Data":"81a4d13585e59e0daf7ddd323ce90c4f906cce621ace2d008358fc587c67ae12"} Mar 21 09:42:02 crc kubenswrapper[4696]: I0321 09:42:02.428077 4696 generic.go:334] "Generic (PLEG): container finished" podID="5aa85936-31bf-4cdb-97fe-fa4e606c3c5f" containerID="0c1c5a946e54d1470b5f35c11d5a3aba97f63b3b86d73900cb78ccf8b0b7f29d" exitCode=0 Mar 21 09:42:02 crc kubenswrapper[4696]: I0321 09:42:02.428139 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568102-pjxb7" event={"ID":"5aa85936-31bf-4cdb-97fe-fa4e606c3c5f","Type":"ContainerDied","Data":"0c1c5a946e54d1470b5f35c11d5a3aba97f63b3b86d73900cb78ccf8b0b7f29d"} Mar 21 09:42:02 crc kubenswrapper[4696]: I0321 09:42:02.535298 4696 scope.go:117] "RemoveContainer" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" Mar 21 09:42:02 crc kubenswrapper[4696]: E0321 09:42:02.535518 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:42:03 crc kubenswrapper[4696]: I0321 09:42:03.809409 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568102-pjxb7" Mar 21 09:42:03 crc kubenswrapper[4696]: I0321 09:42:03.889277 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnjk4\" (UniqueName: \"kubernetes.io/projected/5aa85936-31bf-4cdb-97fe-fa4e606c3c5f-kube-api-access-mnjk4\") pod \"5aa85936-31bf-4cdb-97fe-fa4e606c3c5f\" (UID: \"5aa85936-31bf-4cdb-97fe-fa4e606c3c5f\") " Mar 21 09:42:03 crc kubenswrapper[4696]: I0321 09:42:03.894636 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5aa85936-31bf-4cdb-97fe-fa4e606c3c5f-kube-api-access-mnjk4" (OuterVolumeSpecName: "kube-api-access-mnjk4") pod "5aa85936-31bf-4cdb-97fe-fa4e606c3c5f" (UID: "5aa85936-31bf-4cdb-97fe-fa4e606c3c5f"). InnerVolumeSpecName "kube-api-access-mnjk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:42:03 crc kubenswrapper[4696]: I0321 09:42:03.991249 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnjk4\" (UniqueName: \"kubernetes.io/projected/5aa85936-31bf-4cdb-97fe-fa4e606c3c5f-kube-api-access-mnjk4\") on node \"crc\" DevicePath \"\"" Mar 21 09:42:04 crc kubenswrapper[4696]: I0321 09:42:04.450382 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568102-pjxb7" event={"ID":"5aa85936-31bf-4cdb-97fe-fa4e606c3c5f","Type":"ContainerDied","Data":"81a4d13585e59e0daf7ddd323ce90c4f906cce621ace2d008358fc587c67ae12"} Mar 21 09:42:04 crc kubenswrapper[4696]: I0321 09:42:04.450419 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81a4d13585e59e0daf7ddd323ce90c4f906cce621ace2d008358fc587c67ae12" Mar 21 09:42:04 crc kubenswrapper[4696]: I0321 09:42:04.450463 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568102-pjxb7" Mar 21 09:42:04 crc kubenswrapper[4696]: I0321 09:42:04.880846 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568096-hq7ss"] Mar 21 09:42:04 crc kubenswrapper[4696]: I0321 09:42:04.890065 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568096-hq7ss"] Mar 21 09:42:06 crc kubenswrapper[4696]: I0321 09:42:06.546879 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73788553-329c-410c-bf8a-549ff702848d" path="/var/lib/kubelet/pods/73788553-329c-410c-bf8a-549ff702848d/volumes" Mar 21 09:42:15 crc kubenswrapper[4696]: I0321 09:42:15.535234 4696 scope.go:117] "RemoveContainer" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" Mar 21 09:42:15 crc kubenswrapper[4696]: E0321 09:42:15.536017 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:42:30 crc kubenswrapper[4696]: I0321 09:42:30.534601 4696 scope.go:117] "RemoveContainer" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" Mar 21 09:42:30 crc kubenswrapper[4696]: E0321 09:42:30.535478 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:42:44 crc kubenswrapper[4696]: I0321 09:42:44.535555 4696 scope.go:117] "RemoveContainer" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" Mar 21 09:42:44 crc kubenswrapper[4696]: E0321 09:42:44.536384 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:42:59 crc kubenswrapper[4696]: I0321 09:42:59.534375 4696 scope.go:117] "RemoveContainer" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" Mar 21 09:42:59 crc kubenswrapper[4696]: E0321 09:42:59.534988 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:43:04 crc kubenswrapper[4696]: I0321 09:43:04.401562 4696 scope.go:117] "RemoveContainer" containerID="f168e3e0b46f425a9f90203de70925dc5494348a1f153d0ad4a5d9f5eb135eb2" Mar 21 09:43:13 crc kubenswrapper[4696]: I0321 09:43:13.535454 4696 scope.go:117] "RemoveContainer" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" Mar 21 09:43:13 crc kubenswrapper[4696]: E0321 09:43:13.536349 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:43:26 crc kubenswrapper[4696]: I0321 09:43:26.547880 4696 scope.go:117] "RemoveContainer" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" Mar 21 09:43:26 crc kubenswrapper[4696]: E0321 09:43:26.548619 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:43:39 crc kubenswrapper[4696]: I0321 09:43:39.535592 4696 scope.go:117] "RemoveContainer" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" Mar 21 09:43:39 crc kubenswrapper[4696]: E0321 09:43:39.536512 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:43:50 crc kubenswrapper[4696]: I0321 09:43:50.535317 4696 scope.go:117] "RemoveContainer" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" Mar 21 09:43:50 crc kubenswrapper[4696]: E0321 09:43:50.536111 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:44:00 crc kubenswrapper[4696]: I0321 09:44:00.149832 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29568104-dlzcc"] Mar 21 09:44:00 crc kubenswrapper[4696]: E0321 09:44:00.151064 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aa85936-31bf-4cdb-97fe-fa4e606c3c5f" containerName="oc" Mar 21 09:44:00 crc kubenswrapper[4696]: I0321 09:44:00.151081 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aa85936-31bf-4cdb-97fe-fa4e606c3c5f" containerName="oc" Mar 21 09:44:00 crc kubenswrapper[4696]: I0321 09:44:00.151402 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="5aa85936-31bf-4cdb-97fe-fa4e606c3c5f" containerName="oc" Mar 21 09:44:00 crc kubenswrapper[4696]: I0321 09:44:00.152450 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568104-dlzcc" Mar 21 09:44:00 crc kubenswrapper[4696]: I0321 09:44:00.155275 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-lqx65" Mar 21 09:44:00 crc kubenswrapper[4696]: I0321 09:44:00.155608 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 21 09:44:00 crc kubenswrapper[4696]: I0321 09:44:00.166201 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 21 09:44:00 crc kubenswrapper[4696]: I0321 09:44:00.166975 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568104-dlzcc"] Mar 21 09:44:00 crc kubenswrapper[4696]: I0321 09:44:00.315124 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdvwk\" (UniqueName: \"kubernetes.io/projected/69f8bdda-7124-4d9f-b8a8-e6fe7509436a-kube-api-access-qdvwk\") pod \"auto-csr-approver-29568104-dlzcc\" (UID: \"69f8bdda-7124-4d9f-b8a8-e6fe7509436a\") " pod="openshift-infra/auto-csr-approver-29568104-dlzcc" Mar 21 09:44:00 crc kubenswrapper[4696]: I0321 09:44:00.417805 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdvwk\" (UniqueName: \"kubernetes.io/projected/69f8bdda-7124-4d9f-b8a8-e6fe7509436a-kube-api-access-qdvwk\") pod \"auto-csr-approver-29568104-dlzcc\" (UID: \"69f8bdda-7124-4d9f-b8a8-e6fe7509436a\") " pod="openshift-infra/auto-csr-approver-29568104-dlzcc" Mar 21 09:44:00 crc kubenswrapper[4696]: I0321 09:44:00.441088 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdvwk\" (UniqueName: \"kubernetes.io/projected/69f8bdda-7124-4d9f-b8a8-e6fe7509436a-kube-api-access-qdvwk\") pod \"auto-csr-approver-29568104-dlzcc\" (UID: \"69f8bdda-7124-4d9f-b8a8-e6fe7509436a\") " pod="openshift-infra/auto-csr-approver-29568104-dlzcc" Mar 21 09:44:00 crc kubenswrapper[4696]: I0321 09:44:00.477285 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568104-dlzcc" Mar 21 09:44:01 crc kubenswrapper[4696]: I0321 09:44:01.019957 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29568104-dlzcc"] Mar 21 09:44:01 crc kubenswrapper[4696]: I0321 09:44:01.536959 4696 scope.go:117] "RemoveContainer" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" Mar 21 09:44:01 crc kubenswrapper[4696]: E0321 09:44:01.538327 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:44:01 crc kubenswrapper[4696]: I0321 09:44:01.654691 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568104-dlzcc" event={"ID":"69f8bdda-7124-4d9f-b8a8-e6fe7509436a","Type":"ContainerStarted","Data":"4faa7acc591cc698d1dab7f2eccbf4693e9e77a731d64f3bafba1ec725640ea1"} Mar 21 09:44:02 crc kubenswrapper[4696]: I0321 09:44:02.703011 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568104-dlzcc" event={"ID":"69f8bdda-7124-4d9f-b8a8-e6fe7509436a","Type":"ContainerStarted","Data":"109eac2a8248564872dc4fc2b50f0a8cf5d1ca2f5c4913f2081d31b2ab7c22d0"} Mar 21 09:44:02 crc kubenswrapper[4696]: I0321 09:44:02.736972 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29568104-dlzcc" podStartSLOduration=1.9427928620000001 podStartE2EDuration="2.736941934s" podCreationTimestamp="2026-03-21 09:44:00 +0000 UTC" firstStartedPulling="2026-03-21 09:44:01.023475587 +0000 UTC m=+4575.144356300" lastFinishedPulling="2026-03-21 09:44:01.817624659 +0000 UTC m=+4575.938505372" observedRunningTime="2026-03-21 09:44:02.723619925 +0000 UTC m=+4576.844500638" watchObservedRunningTime="2026-03-21 09:44:02.736941934 +0000 UTC m=+4576.857822647" Mar 21 09:44:03 crc kubenswrapper[4696]: I0321 09:44:03.721644 4696 generic.go:334] "Generic (PLEG): container finished" podID="69f8bdda-7124-4d9f-b8a8-e6fe7509436a" containerID="109eac2a8248564872dc4fc2b50f0a8cf5d1ca2f5c4913f2081d31b2ab7c22d0" exitCode=0 Mar 21 09:44:03 crc kubenswrapper[4696]: I0321 09:44:03.723788 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568104-dlzcc" event={"ID":"69f8bdda-7124-4d9f-b8a8-e6fe7509436a","Type":"ContainerDied","Data":"109eac2a8248564872dc4fc2b50f0a8cf5d1ca2f5c4913f2081d31b2ab7c22d0"} Mar 21 09:44:05 crc kubenswrapper[4696]: I0321 09:44:05.119615 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568104-dlzcc" Mar 21 09:44:05 crc kubenswrapper[4696]: I0321 09:44:05.254218 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdvwk\" (UniqueName: \"kubernetes.io/projected/69f8bdda-7124-4d9f-b8a8-e6fe7509436a-kube-api-access-qdvwk\") pod \"69f8bdda-7124-4d9f-b8a8-e6fe7509436a\" (UID: \"69f8bdda-7124-4d9f-b8a8-e6fe7509436a\") " Mar 21 09:44:05 crc kubenswrapper[4696]: I0321 09:44:05.259890 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69f8bdda-7124-4d9f-b8a8-e6fe7509436a-kube-api-access-qdvwk" (OuterVolumeSpecName: "kube-api-access-qdvwk") pod "69f8bdda-7124-4d9f-b8a8-e6fe7509436a" (UID: "69f8bdda-7124-4d9f-b8a8-e6fe7509436a"). InnerVolumeSpecName "kube-api-access-qdvwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:44:05 crc kubenswrapper[4696]: I0321 09:44:05.356460 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdvwk\" (UniqueName: \"kubernetes.io/projected/69f8bdda-7124-4d9f-b8a8-e6fe7509436a-kube-api-access-qdvwk\") on node \"crc\" DevicePath \"\"" Mar 21 09:44:05 crc kubenswrapper[4696]: I0321 09:44:05.751499 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29568104-dlzcc" event={"ID":"69f8bdda-7124-4d9f-b8a8-e6fe7509436a","Type":"ContainerDied","Data":"4faa7acc591cc698d1dab7f2eccbf4693e9e77a731d64f3bafba1ec725640ea1"} Mar 21 09:44:05 crc kubenswrapper[4696]: I0321 09:44:05.751580 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4faa7acc591cc698d1dab7f2eccbf4693e9e77a731d64f3bafba1ec725640ea1" Mar 21 09:44:05 crc kubenswrapper[4696]: I0321 09:44:05.751689 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29568104-dlzcc" Mar 21 09:44:05 crc kubenswrapper[4696]: I0321 09:44:05.811870 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29568098-zct9z"] Mar 21 09:44:05 crc kubenswrapper[4696]: I0321 09:44:05.822221 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29568098-zct9z"] Mar 21 09:44:06 crc kubenswrapper[4696]: I0321 09:44:06.547662 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f3b9dba-3a21-4101-92c3-59b62496539a" path="/var/lib/kubelet/pods/6f3b9dba-3a21-4101-92c3-59b62496539a/volumes" Mar 21 09:44:14 crc kubenswrapper[4696]: I0321 09:44:14.536133 4696 scope.go:117] "RemoveContainer" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" Mar 21 09:44:14 crc kubenswrapper[4696]: E0321 09:44:14.537130 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:44:19 crc kubenswrapper[4696]: I0321 09:44:19.745191 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hs8bx"] Mar 21 09:44:19 crc kubenswrapper[4696]: E0321 09:44:19.746101 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69f8bdda-7124-4d9f-b8a8-e6fe7509436a" containerName="oc" Mar 21 09:44:19 crc kubenswrapper[4696]: I0321 09:44:19.746117 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="69f8bdda-7124-4d9f-b8a8-e6fe7509436a" containerName="oc" Mar 21 09:44:19 crc kubenswrapper[4696]: I0321 09:44:19.746385 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="69f8bdda-7124-4d9f-b8a8-e6fe7509436a" containerName="oc" Mar 21 09:44:19 crc kubenswrapper[4696]: I0321 09:44:19.748332 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hs8bx" Mar 21 09:44:19 crc kubenswrapper[4696]: I0321 09:44:19.794403 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hs8bx"] Mar 21 09:44:19 crc kubenswrapper[4696]: I0321 09:44:19.853150 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4488c5a5-4d4d-44d3-a734-ac502836ee01-utilities\") pod \"redhat-marketplace-hs8bx\" (UID: \"4488c5a5-4d4d-44d3-a734-ac502836ee01\") " pod="openshift-marketplace/redhat-marketplace-hs8bx" Mar 21 09:44:19 crc kubenswrapper[4696]: I0321 09:44:19.853291 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnztd\" (UniqueName: \"kubernetes.io/projected/4488c5a5-4d4d-44d3-a734-ac502836ee01-kube-api-access-mnztd\") pod \"redhat-marketplace-hs8bx\" (UID: \"4488c5a5-4d4d-44d3-a734-ac502836ee01\") " pod="openshift-marketplace/redhat-marketplace-hs8bx" Mar 21 09:44:19 crc kubenswrapper[4696]: I0321 09:44:19.853393 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4488c5a5-4d4d-44d3-a734-ac502836ee01-catalog-content\") pod \"redhat-marketplace-hs8bx\" (UID: \"4488c5a5-4d4d-44d3-a734-ac502836ee01\") " pod="openshift-marketplace/redhat-marketplace-hs8bx" Mar 21 09:44:19 crc kubenswrapper[4696]: I0321 09:44:19.954642 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnztd\" (UniqueName: \"kubernetes.io/projected/4488c5a5-4d4d-44d3-a734-ac502836ee01-kube-api-access-mnztd\") pod \"redhat-marketplace-hs8bx\" (UID: \"4488c5a5-4d4d-44d3-a734-ac502836ee01\") " pod="openshift-marketplace/redhat-marketplace-hs8bx" Mar 21 09:44:19 crc kubenswrapper[4696]: I0321 09:44:19.954792 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4488c5a5-4d4d-44d3-a734-ac502836ee01-catalog-content\") pod \"redhat-marketplace-hs8bx\" (UID: \"4488c5a5-4d4d-44d3-a734-ac502836ee01\") " pod="openshift-marketplace/redhat-marketplace-hs8bx" Mar 21 09:44:19 crc kubenswrapper[4696]: I0321 09:44:19.954877 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4488c5a5-4d4d-44d3-a734-ac502836ee01-utilities\") pod \"redhat-marketplace-hs8bx\" (UID: \"4488c5a5-4d4d-44d3-a734-ac502836ee01\") " pod="openshift-marketplace/redhat-marketplace-hs8bx" Mar 21 09:44:19 crc kubenswrapper[4696]: I0321 09:44:19.955280 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4488c5a5-4d4d-44d3-a734-ac502836ee01-catalog-content\") pod \"redhat-marketplace-hs8bx\" (UID: \"4488c5a5-4d4d-44d3-a734-ac502836ee01\") " pod="openshift-marketplace/redhat-marketplace-hs8bx" Mar 21 09:44:19 crc kubenswrapper[4696]: I0321 09:44:19.955434 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4488c5a5-4d4d-44d3-a734-ac502836ee01-utilities\") pod \"redhat-marketplace-hs8bx\" (UID: \"4488c5a5-4d4d-44d3-a734-ac502836ee01\") " pod="openshift-marketplace/redhat-marketplace-hs8bx" Mar 21 09:44:19 crc kubenswrapper[4696]: I0321 09:44:19.995109 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnztd\" (UniqueName: \"kubernetes.io/projected/4488c5a5-4d4d-44d3-a734-ac502836ee01-kube-api-access-mnztd\") pod \"redhat-marketplace-hs8bx\" (UID: \"4488c5a5-4d4d-44d3-a734-ac502836ee01\") " pod="openshift-marketplace/redhat-marketplace-hs8bx" Mar 21 09:44:20 crc kubenswrapper[4696]: I0321 09:44:20.078435 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hs8bx" Mar 21 09:44:20 crc kubenswrapper[4696]: I0321 09:44:20.546872 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hs8bx"] Mar 21 09:44:20 crc kubenswrapper[4696]: I0321 09:44:20.895504 4696 generic.go:334] "Generic (PLEG): container finished" podID="4488c5a5-4d4d-44d3-a734-ac502836ee01" containerID="d8d0370d93becd99f3a86e9b9779d704f3a1afe1498cabc35eff9b8be9601154" exitCode=0 Mar 21 09:44:20 crc kubenswrapper[4696]: I0321 09:44:20.895725 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hs8bx" event={"ID":"4488c5a5-4d4d-44d3-a734-ac502836ee01","Type":"ContainerDied","Data":"d8d0370d93becd99f3a86e9b9779d704f3a1afe1498cabc35eff9b8be9601154"} Mar 21 09:44:20 crc kubenswrapper[4696]: I0321 09:44:20.895791 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hs8bx" event={"ID":"4488c5a5-4d4d-44d3-a734-ac502836ee01","Type":"ContainerStarted","Data":"045b5e0db9704d92e3640c82b9fa64949482ff23dc4b40887d57b948bbbd7ea5"} Mar 21 09:44:22 crc kubenswrapper[4696]: I0321 09:44:22.917412 4696 generic.go:334] "Generic (PLEG): container finished" podID="4488c5a5-4d4d-44d3-a734-ac502836ee01" containerID="600e732fb557786091eec9e9791b2fcce15d6666a4433aef8dc4bd57696896f2" exitCode=0 Mar 21 09:44:22 crc kubenswrapper[4696]: I0321 09:44:22.917990 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hs8bx" event={"ID":"4488c5a5-4d4d-44d3-a734-ac502836ee01","Type":"ContainerDied","Data":"600e732fb557786091eec9e9791b2fcce15d6666a4433aef8dc4bd57696896f2"} Mar 21 09:44:24 crc kubenswrapper[4696]: I0321 09:44:24.940516 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hs8bx" event={"ID":"4488c5a5-4d4d-44d3-a734-ac502836ee01","Type":"ContainerStarted","Data":"e2d4f06d6e33ee364441eafe2e8fab9a2b7ca27188675ad0bbead990aca91b0c"} Mar 21 09:44:24 crc kubenswrapper[4696]: I0321 09:44:24.969967 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hs8bx" podStartSLOduration=3.524332843 podStartE2EDuration="5.969945613s" podCreationTimestamp="2026-03-21 09:44:19 +0000 UTC" firstStartedPulling="2026-03-21 09:44:20.897076553 +0000 UTC m=+4595.017957266" lastFinishedPulling="2026-03-21 09:44:23.342689323 +0000 UTC m=+4597.463570036" observedRunningTime="2026-03-21 09:44:24.964698004 +0000 UTC m=+4599.085578717" watchObservedRunningTime="2026-03-21 09:44:24.969945613 +0000 UTC m=+4599.090826326" Mar 21 09:44:26 crc kubenswrapper[4696]: I0321 09:44:26.543304 4696 scope.go:117] "RemoveContainer" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" Mar 21 09:44:26 crc kubenswrapper[4696]: E0321 09:44:26.544960 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:44:30 crc kubenswrapper[4696]: I0321 09:44:30.078584 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hs8bx" Mar 21 09:44:30 crc kubenswrapper[4696]: I0321 09:44:30.079379 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hs8bx" Mar 21 09:44:30 crc kubenswrapper[4696]: I0321 09:44:30.128272 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hs8bx" Mar 21 09:44:30 crc kubenswrapper[4696]: I0321 09:44:30.378502 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kjxk5"] Mar 21 09:44:30 crc kubenswrapper[4696]: I0321 09:44:30.381225 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kjxk5" Mar 21 09:44:30 crc kubenswrapper[4696]: I0321 09:44:30.392802 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kjxk5"] Mar 21 09:44:30 crc kubenswrapper[4696]: I0321 09:44:30.490574 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljpkk\" (UniqueName: \"kubernetes.io/projected/38321a76-5396-41dc-8595-8c672ead248e-kube-api-access-ljpkk\") pod \"certified-operators-kjxk5\" (UID: \"38321a76-5396-41dc-8595-8c672ead248e\") " pod="openshift-marketplace/certified-operators-kjxk5" Mar 21 09:44:30 crc kubenswrapper[4696]: I0321 09:44:30.490648 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38321a76-5396-41dc-8595-8c672ead248e-catalog-content\") pod \"certified-operators-kjxk5\" (UID: \"38321a76-5396-41dc-8595-8c672ead248e\") " pod="openshift-marketplace/certified-operators-kjxk5" Mar 21 09:44:30 crc kubenswrapper[4696]: I0321 09:44:30.490691 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38321a76-5396-41dc-8595-8c672ead248e-utilities\") pod \"certified-operators-kjxk5\" (UID: \"38321a76-5396-41dc-8595-8c672ead248e\") " pod="openshift-marketplace/certified-operators-kjxk5" Mar 21 09:44:30 crc kubenswrapper[4696]: I0321 09:44:30.592861 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljpkk\" (UniqueName: \"kubernetes.io/projected/38321a76-5396-41dc-8595-8c672ead248e-kube-api-access-ljpkk\") pod \"certified-operators-kjxk5\" (UID: \"38321a76-5396-41dc-8595-8c672ead248e\") " pod="openshift-marketplace/certified-operators-kjxk5" Mar 21 09:44:30 crc kubenswrapper[4696]: I0321 09:44:30.593282 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38321a76-5396-41dc-8595-8c672ead248e-catalog-content\") pod \"certified-operators-kjxk5\" (UID: \"38321a76-5396-41dc-8595-8c672ead248e\") " pod="openshift-marketplace/certified-operators-kjxk5" Mar 21 09:44:30 crc kubenswrapper[4696]: I0321 09:44:30.593333 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38321a76-5396-41dc-8595-8c672ead248e-utilities\") pod \"certified-operators-kjxk5\" (UID: \"38321a76-5396-41dc-8595-8c672ead248e\") " pod="openshift-marketplace/certified-operators-kjxk5" Mar 21 09:44:30 crc kubenswrapper[4696]: I0321 09:44:30.593993 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38321a76-5396-41dc-8595-8c672ead248e-utilities\") pod \"certified-operators-kjxk5\" (UID: \"38321a76-5396-41dc-8595-8c672ead248e\") " pod="openshift-marketplace/certified-operators-kjxk5" Mar 21 09:44:30 crc kubenswrapper[4696]: I0321 09:44:30.594585 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38321a76-5396-41dc-8595-8c672ead248e-catalog-content\") pod \"certified-operators-kjxk5\" (UID: \"38321a76-5396-41dc-8595-8c672ead248e\") " pod="openshift-marketplace/certified-operators-kjxk5" Mar 21 09:44:30 crc kubenswrapper[4696]: I0321 09:44:30.614231 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljpkk\" (UniqueName: \"kubernetes.io/projected/38321a76-5396-41dc-8595-8c672ead248e-kube-api-access-ljpkk\") pod \"certified-operators-kjxk5\" (UID: \"38321a76-5396-41dc-8595-8c672ead248e\") " pod="openshift-marketplace/certified-operators-kjxk5" Mar 21 09:44:30 crc kubenswrapper[4696]: I0321 09:44:30.711611 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kjxk5" Mar 21 09:44:31 crc kubenswrapper[4696]: I0321 09:44:31.108952 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hs8bx" Mar 21 09:44:31 crc kubenswrapper[4696]: I0321 09:44:31.319864 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kjxk5"] Mar 21 09:44:32 crc kubenswrapper[4696]: I0321 09:44:32.023926 4696 generic.go:334] "Generic (PLEG): container finished" podID="38321a76-5396-41dc-8595-8c672ead248e" containerID="73e089df40dc7ac88610b7854184447b73c23dc7b8bf578ae045d2f3f3395bb2" exitCode=0 Mar 21 09:44:32 crc kubenswrapper[4696]: I0321 09:44:32.024082 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kjxk5" event={"ID":"38321a76-5396-41dc-8595-8c672ead248e","Type":"ContainerDied","Data":"73e089df40dc7ac88610b7854184447b73c23dc7b8bf578ae045d2f3f3395bb2"} Mar 21 09:44:32 crc kubenswrapper[4696]: I0321 09:44:32.025076 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kjxk5" event={"ID":"38321a76-5396-41dc-8595-8c672ead248e","Type":"ContainerStarted","Data":"4a4c4e9f6d2147880a8f265fea700cd55c6bc75a9eaaec458066632c0ec751cc"} Mar 21 09:44:33 crc kubenswrapper[4696]: I0321 09:44:33.364615 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hs8bx"] Mar 21 09:44:33 crc kubenswrapper[4696]: I0321 09:44:33.365335 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hs8bx" podUID="4488c5a5-4d4d-44d3-a734-ac502836ee01" containerName="registry-server" containerID="cri-o://e2d4f06d6e33ee364441eafe2e8fab9a2b7ca27188675ad0bbead990aca91b0c" gracePeriod=2 Mar 21 09:44:34 crc kubenswrapper[4696]: I0321 09:44:34.050946 4696 generic.go:334] "Generic (PLEG): container finished" podID="4488c5a5-4d4d-44d3-a734-ac502836ee01" containerID="e2d4f06d6e33ee364441eafe2e8fab9a2b7ca27188675ad0bbead990aca91b0c" exitCode=0 Mar 21 09:44:34 crc kubenswrapper[4696]: I0321 09:44:34.051006 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hs8bx" event={"ID":"4488c5a5-4d4d-44d3-a734-ac502836ee01","Type":"ContainerDied","Data":"e2d4f06d6e33ee364441eafe2e8fab9a2b7ca27188675ad0bbead990aca91b0c"} Mar 21 09:44:34 crc kubenswrapper[4696]: I0321 09:44:34.051032 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hs8bx" event={"ID":"4488c5a5-4d4d-44d3-a734-ac502836ee01","Type":"ContainerDied","Data":"045b5e0db9704d92e3640c82b9fa64949482ff23dc4b40887d57b948bbbd7ea5"} Mar 21 09:44:34 crc kubenswrapper[4696]: I0321 09:44:34.051043 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="045b5e0db9704d92e3640c82b9fa64949482ff23dc4b40887d57b948bbbd7ea5" Mar 21 09:44:34 crc kubenswrapper[4696]: I0321 09:44:34.053284 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kjxk5" event={"ID":"38321a76-5396-41dc-8595-8c672ead248e","Type":"ContainerStarted","Data":"200e223d63efcb0d9ed3e1868f0b6ba4becc7e20a5c0871ae870c7f3d9148248"} Mar 21 09:44:34 crc kubenswrapper[4696]: I0321 09:44:34.640526 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hs8bx" Mar 21 09:44:34 crc kubenswrapper[4696]: I0321 09:44:34.724551 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnztd\" (UniqueName: \"kubernetes.io/projected/4488c5a5-4d4d-44d3-a734-ac502836ee01-kube-api-access-mnztd\") pod \"4488c5a5-4d4d-44d3-a734-ac502836ee01\" (UID: \"4488c5a5-4d4d-44d3-a734-ac502836ee01\") " Mar 21 09:44:34 crc kubenswrapper[4696]: I0321 09:44:34.724703 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4488c5a5-4d4d-44d3-a734-ac502836ee01-utilities\") pod \"4488c5a5-4d4d-44d3-a734-ac502836ee01\" (UID: \"4488c5a5-4d4d-44d3-a734-ac502836ee01\") " Mar 21 09:44:34 crc kubenswrapper[4696]: I0321 09:44:34.724782 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4488c5a5-4d4d-44d3-a734-ac502836ee01-catalog-content\") pod \"4488c5a5-4d4d-44d3-a734-ac502836ee01\" (UID: \"4488c5a5-4d4d-44d3-a734-ac502836ee01\") " Mar 21 09:44:34 crc kubenswrapper[4696]: I0321 09:44:34.726029 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4488c5a5-4d4d-44d3-a734-ac502836ee01-utilities" (OuterVolumeSpecName: "utilities") pod "4488c5a5-4d4d-44d3-a734-ac502836ee01" (UID: "4488c5a5-4d4d-44d3-a734-ac502836ee01"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:44:34 crc kubenswrapper[4696]: I0321 09:44:34.735617 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4488c5a5-4d4d-44d3-a734-ac502836ee01-kube-api-access-mnztd" (OuterVolumeSpecName: "kube-api-access-mnztd") pod "4488c5a5-4d4d-44d3-a734-ac502836ee01" (UID: "4488c5a5-4d4d-44d3-a734-ac502836ee01"). InnerVolumeSpecName "kube-api-access-mnztd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:44:34 crc kubenswrapper[4696]: I0321 09:44:34.754338 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4488c5a5-4d4d-44d3-a734-ac502836ee01-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4488c5a5-4d4d-44d3-a734-ac502836ee01" (UID: "4488c5a5-4d4d-44d3-a734-ac502836ee01"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:44:34 crc kubenswrapper[4696]: I0321 09:44:34.827785 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnztd\" (UniqueName: \"kubernetes.io/projected/4488c5a5-4d4d-44d3-a734-ac502836ee01-kube-api-access-mnztd\") on node \"crc\" DevicePath \"\"" Mar 21 09:44:34 crc kubenswrapper[4696]: I0321 09:44:34.828042 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4488c5a5-4d4d-44d3-a734-ac502836ee01-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 09:44:34 crc kubenswrapper[4696]: I0321 09:44:34.828053 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4488c5a5-4d4d-44d3-a734-ac502836ee01-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 09:44:35 crc kubenswrapper[4696]: I0321 09:44:35.067390 4696 generic.go:334] "Generic (PLEG): container finished" podID="38321a76-5396-41dc-8595-8c672ead248e" containerID="200e223d63efcb0d9ed3e1868f0b6ba4becc7e20a5c0871ae870c7f3d9148248" exitCode=0 Mar 21 09:44:35 crc kubenswrapper[4696]: I0321 09:44:35.067487 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hs8bx" Mar 21 09:44:35 crc kubenswrapper[4696]: I0321 09:44:35.068863 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kjxk5" event={"ID":"38321a76-5396-41dc-8595-8c672ead248e","Type":"ContainerDied","Data":"200e223d63efcb0d9ed3e1868f0b6ba4becc7e20a5c0871ae870c7f3d9148248"} Mar 21 09:44:35 crc kubenswrapper[4696]: I0321 09:44:35.111283 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hs8bx"] Mar 21 09:44:35 crc kubenswrapper[4696]: I0321 09:44:35.119537 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hs8bx"] Mar 21 09:44:36 crc kubenswrapper[4696]: I0321 09:44:36.106934 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kjxk5" event={"ID":"38321a76-5396-41dc-8595-8c672ead248e","Type":"ContainerStarted","Data":"c7138436f2b46239aab758c8464204b1bac00bba0fc35dcc881af203a379fde4"} Mar 21 09:44:36 crc kubenswrapper[4696]: I0321 09:44:36.132504 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kjxk5" podStartSLOduration=2.704162501 podStartE2EDuration="6.132484434s" podCreationTimestamp="2026-03-21 09:44:30 +0000 UTC" firstStartedPulling="2026-03-21 09:44:32.026577328 +0000 UTC m=+4606.147458041" lastFinishedPulling="2026-03-21 09:44:35.454899241 +0000 UTC m=+4609.575779974" observedRunningTime="2026-03-21 09:44:36.131406043 +0000 UTC m=+4610.252286756" watchObservedRunningTime="2026-03-21 09:44:36.132484434 +0000 UTC m=+4610.253365147" Mar 21 09:44:36 crc kubenswrapper[4696]: I0321 09:44:36.545618 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4488c5a5-4d4d-44d3-a734-ac502836ee01" path="/var/lib/kubelet/pods/4488c5a5-4d4d-44d3-a734-ac502836ee01/volumes" Mar 21 09:44:40 crc kubenswrapper[4696]: I0321 09:44:40.535632 4696 scope.go:117] "RemoveContainer" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" Mar 21 09:44:40 crc kubenswrapper[4696]: E0321 09:44:40.536717 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:44:40 crc kubenswrapper[4696]: I0321 09:44:40.711739 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kjxk5" Mar 21 09:44:40 crc kubenswrapper[4696]: I0321 09:44:40.711839 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kjxk5" Mar 21 09:44:40 crc kubenswrapper[4696]: I0321 09:44:40.773437 4696 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kjxk5" Mar 21 09:44:41 crc kubenswrapper[4696]: I0321 09:44:41.200041 4696 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kjxk5" Mar 21 09:44:41 crc kubenswrapper[4696]: I0321 09:44:41.263063 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kjxk5"] Mar 21 09:44:43 crc kubenswrapper[4696]: I0321 09:44:43.172894 4696 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kjxk5" podUID="38321a76-5396-41dc-8595-8c672ead248e" containerName="registry-server" containerID="cri-o://c7138436f2b46239aab758c8464204b1bac00bba0fc35dcc881af203a379fde4" gracePeriod=2 Mar 21 09:44:44 crc kubenswrapper[4696]: I0321 09:44:44.184672 4696 generic.go:334] "Generic (PLEG): container finished" podID="38321a76-5396-41dc-8595-8c672ead248e" containerID="c7138436f2b46239aab758c8464204b1bac00bba0fc35dcc881af203a379fde4" exitCode=0 Mar 21 09:44:44 crc kubenswrapper[4696]: I0321 09:44:44.185291 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kjxk5" event={"ID":"38321a76-5396-41dc-8595-8c672ead248e","Type":"ContainerDied","Data":"c7138436f2b46239aab758c8464204b1bac00bba0fc35dcc881af203a379fde4"} Mar 21 09:44:44 crc kubenswrapper[4696]: I0321 09:44:44.185328 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kjxk5" event={"ID":"38321a76-5396-41dc-8595-8c672ead248e","Type":"ContainerDied","Data":"4a4c4e9f6d2147880a8f265fea700cd55c6bc75a9eaaec458066632c0ec751cc"} Mar 21 09:44:44 crc kubenswrapper[4696]: I0321 09:44:44.185345 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a4c4e9f6d2147880a8f265fea700cd55c6bc75a9eaaec458066632c0ec751cc" Mar 21 09:44:44 crc kubenswrapper[4696]: I0321 09:44:44.529327 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kjxk5" Mar 21 09:44:44 crc kubenswrapper[4696]: I0321 09:44:44.657282 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38321a76-5396-41dc-8595-8c672ead248e-utilities\") pod \"38321a76-5396-41dc-8595-8c672ead248e\" (UID: \"38321a76-5396-41dc-8595-8c672ead248e\") " Mar 21 09:44:44 crc kubenswrapper[4696]: I0321 09:44:44.657690 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38321a76-5396-41dc-8595-8c672ead248e-catalog-content\") pod \"38321a76-5396-41dc-8595-8c672ead248e\" (UID: \"38321a76-5396-41dc-8595-8c672ead248e\") " Mar 21 09:44:44 crc kubenswrapper[4696]: I0321 09:44:44.657919 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljpkk\" (UniqueName: \"kubernetes.io/projected/38321a76-5396-41dc-8595-8c672ead248e-kube-api-access-ljpkk\") pod \"38321a76-5396-41dc-8595-8c672ead248e\" (UID: \"38321a76-5396-41dc-8595-8c672ead248e\") " Mar 21 09:44:44 crc kubenswrapper[4696]: I0321 09:44:44.658219 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38321a76-5396-41dc-8595-8c672ead248e-utilities" (OuterVolumeSpecName: "utilities") pod "38321a76-5396-41dc-8595-8c672ead248e" (UID: "38321a76-5396-41dc-8595-8c672ead248e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:44:44 crc kubenswrapper[4696]: I0321 09:44:44.658982 4696 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38321a76-5396-41dc-8595-8c672ead248e-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 09:44:44 crc kubenswrapper[4696]: I0321 09:44:44.667773 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38321a76-5396-41dc-8595-8c672ead248e-kube-api-access-ljpkk" (OuterVolumeSpecName: "kube-api-access-ljpkk") pod "38321a76-5396-41dc-8595-8c672ead248e" (UID: "38321a76-5396-41dc-8595-8c672ead248e"). InnerVolumeSpecName "kube-api-access-ljpkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:44:44 crc kubenswrapper[4696]: I0321 09:44:44.706999 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38321a76-5396-41dc-8595-8c672ead248e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38321a76-5396-41dc-8595-8c672ead248e" (UID: "38321a76-5396-41dc-8595-8c672ead248e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 21 09:44:44 crc kubenswrapper[4696]: I0321 09:44:44.760784 4696 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38321a76-5396-41dc-8595-8c672ead248e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 09:44:44 crc kubenswrapper[4696]: I0321 09:44:44.760820 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljpkk\" (UniqueName: \"kubernetes.io/projected/38321a76-5396-41dc-8595-8c672ead248e-kube-api-access-ljpkk\") on node \"crc\" DevicePath \"\"" Mar 21 09:44:45 crc kubenswrapper[4696]: I0321 09:44:45.196588 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kjxk5" Mar 21 09:44:45 crc kubenswrapper[4696]: I0321 09:44:45.240571 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kjxk5"] Mar 21 09:44:45 crc kubenswrapper[4696]: I0321 09:44:45.250604 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kjxk5"] Mar 21 09:44:46 crc kubenswrapper[4696]: I0321 09:44:46.546564 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38321a76-5396-41dc-8595-8c672ead248e" path="/var/lib/kubelet/pods/38321a76-5396-41dc-8595-8c672ead248e/volumes" Mar 21 09:44:51 crc kubenswrapper[4696]: I0321 09:44:51.534743 4696 scope.go:117] "RemoveContainer" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" Mar 21 09:44:51 crc kubenswrapper[4696]: E0321 09:44:51.535593 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.157299 4696 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568105-ktz8q"] Mar 21 09:45:00 crc kubenswrapper[4696]: E0321 09:45:00.158543 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38321a76-5396-41dc-8595-8c672ead248e" containerName="registry-server" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.158567 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="38321a76-5396-41dc-8595-8c672ead248e" containerName="registry-server" Mar 21 09:45:00 crc kubenswrapper[4696]: E0321 09:45:00.158604 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4488c5a5-4d4d-44d3-a734-ac502836ee01" containerName="extract-content" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.158616 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4488c5a5-4d4d-44d3-a734-ac502836ee01" containerName="extract-content" Mar 21 09:45:00 crc kubenswrapper[4696]: E0321 09:45:00.158654 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38321a76-5396-41dc-8595-8c672ead248e" containerName="extract-content" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.158665 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="38321a76-5396-41dc-8595-8c672ead248e" containerName="extract-content" Mar 21 09:45:00 crc kubenswrapper[4696]: E0321 09:45:00.158685 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38321a76-5396-41dc-8595-8c672ead248e" containerName="extract-utilities" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.158696 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="38321a76-5396-41dc-8595-8c672ead248e" containerName="extract-utilities" Mar 21 09:45:00 crc kubenswrapper[4696]: E0321 09:45:00.158726 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4488c5a5-4d4d-44d3-a734-ac502836ee01" containerName="registry-server" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.158737 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4488c5a5-4d4d-44d3-a734-ac502836ee01" containerName="registry-server" Mar 21 09:45:00 crc kubenswrapper[4696]: E0321 09:45:00.158759 4696 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4488c5a5-4d4d-44d3-a734-ac502836ee01" containerName="extract-utilities" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.158770 4696 state_mem.go:107] "Deleted CPUSet assignment" podUID="4488c5a5-4d4d-44d3-a734-ac502836ee01" containerName="extract-utilities" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.159123 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="4488c5a5-4d4d-44d3-a734-ac502836ee01" containerName="registry-server" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.159147 4696 memory_manager.go:354] "RemoveStaleState removing state" podUID="38321a76-5396-41dc-8595-8c672ead248e" containerName="registry-server" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.160419 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568105-ktz8q" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.162413 4696 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.162544 4696 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.179249 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568105-ktz8q"] Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.323676 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px2z9\" (UniqueName: \"kubernetes.io/projected/09513529-8f13-4b78-82df-68f14b9e76b5-kube-api-access-px2z9\") pod \"collect-profiles-29568105-ktz8q\" (UID: \"09513529-8f13-4b78-82df-68f14b9e76b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568105-ktz8q" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.323863 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09513529-8f13-4b78-82df-68f14b9e76b5-config-volume\") pod \"collect-profiles-29568105-ktz8q\" (UID: \"09513529-8f13-4b78-82df-68f14b9e76b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568105-ktz8q" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.323943 4696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09513529-8f13-4b78-82df-68f14b9e76b5-secret-volume\") pod \"collect-profiles-29568105-ktz8q\" (UID: \"09513529-8f13-4b78-82df-68f14b9e76b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568105-ktz8q" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.425546 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px2z9\" (UniqueName: \"kubernetes.io/projected/09513529-8f13-4b78-82df-68f14b9e76b5-kube-api-access-px2z9\") pod \"collect-profiles-29568105-ktz8q\" (UID: \"09513529-8f13-4b78-82df-68f14b9e76b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568105-ktz8q" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.426177 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09513529-8f13-4b78-82df-68f14b9e76b5-config-volume\") pod \"collect-profiles-29568105-ktz8q\" (UID: \"09513529-8f13-4b78-82df-68f14b9e76b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568105-ktz8q" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.427247 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09513529-8f13-4b78-82df-68f14b9e76b5-config-volume\") pod \"collect-profiles-29568105-ktz8q\" (UID: \"09513529-8f13-4b78-82df-68f14b9e76b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568105-ktz8q" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.427326 4696 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09513529-8f13-4b78-82df-68f14b9e76b5-secret-volume\") pod \"collect-profiles-29568105-ktz8q\" (UID: \"09513529-8f13-4b78-82df-68f14b9e76b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568105-ktz8q" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.436410 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09513529-8f13-4b78-82df-68f14b9e76b5-secret-volume\") pod \"collect-profiles-29568105-ktz8q\" (UID: \"09513529-8f13-4b78-82df-68f14b9e76b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568105-ktz8q" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.446270 4696 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px2z9\" (UniqueName: \"kubernetes.io/projected/09513529-8f13-4b78-82df-68f14b9e76b5-kube-api-access-px2z9\") pod \"collect-profiles-29568105-ktz8q\" (UID: \"09513529-8f13-4b78-82df-68f14b9e76b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29568105-ktz8q" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.486554 4696 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568105-ktz8q" Mar 21 09:45:00 crc kubenswrapper[4696]: I0321 09:45:00.918370 4696 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568105-ktz8q"] Mar 21 09:45:01 crc kubenswrapper[4696]: I0321 09:45:01.349848 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568105-ktz8q" event={"ID":"09513529-8f13-4b78-82df-68f14b9e76b5","Type":"ContainerStarted","Data":"7e1d04df42c81acfd8f7104e6832265308b8a8a2054106d22bc9520738b12130"} Mar 21 09:45:01 crc kubenswrapper[4696]: I0321 09:45:01.350169 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568105-ktz8q" event={"ID":"09513529-8f13-4b78-82df-68f14b9e76b5","Type":"ContainerStarted","Data":"6ab49f60a1f4202de2276703548b4bf3dc1679e1b52e6c338a860bf8eec5bc16"} Mar 21 09:45:01 crc kubenswrapper[4696]: I0321 09:45:01.367771 4696 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29568105-ktz8q" podStartSLOduration=1.367749152 podStartE2EDuration="1.367749152s" podCreationTimestamp="2026-03-21 09:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 09:45:01.366554079 +0000 UTC m=+4635.487434802" watchObservedRunningTime="2026-03-21 09:45:01.367749152 +0000 UTC m=+4635.488629865" Mar 21 09:45:02 crc kubenswrapper[4696]: I0321 09:45:02.359277 4696 generic.go:334] "Generic (PLEG): container finished" podID="09513529-8f13-4b78-82df-68f14b9e76b5" containerID="7e1d04df42c81acfd8f7104e6832265308b8a8a2054106d22bc9520738b12130" exitCode=0 Mar 21 09:45:02 crc kubenswrapper[4696]: I0321 09:45:02.359330 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568105-ktz8q" event={"ID":"09513529-8f13-4b78-82df-68f14b9e76b5","Type":"ContainerDied","Data":"7e1d04df42c81acfd8f7104e6832265308b8a8a2054106d22bc9520738b12130"} Mar 21 09:45:03 crc kubenswrapper[4696]: I0321 09:45:03.853472 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568105-ktz8q" Mar 21 09:45:04 crc kubenswrapper[4696]: I0321 09:45:04.000390 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09513529-8f13-4b78-82df-68f14b9e76b5-secret-volume\") pod \"09513529-8f13-4b78-82df-68f14b9e76b5\" (UID: \"09513529-8f13-4b78-82df-68f14b9e76b5\") " Mar 21 09:45:04 crc kubenswrapper[4696]: I0321 09:45:04.000452 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px2z9\" (UniqueName: \"kubernetes.io/projected/09513529-8f13-4b78-82df-68f14b9e76b5-kube-api-access-px2z9\") pod \"09513529-8f13-4b78-82df-68f14b9e76b5\" (UID: \"09513529-8f13-4b78-82df-68f14b9e76b5\") " Mar 21 09:45:04 crc kubenswrapper[4696]: I0321 09:45:04.000525 4696 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09513529-8f13-4b78-82df-68f14b9e76b5-config-volume\") pod \"09513529-8f13-4b78-82df-68f14b9e76b5\" (UID: \"09513529-8f13-4b78-82df-68f14b9e76b5\") " Mar 21 09:45:04 crc kubenswrapper[4696]: I0321 09:45:04.001626 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09513529-8f13-4b78-82df-68f14b9e76b5-config-volume" (OuterVolumeSpecName: "config-volume") pod "09513529-8f13-4b78-82df-68f14b9e76b5" (UID: "09513529-8f13-4b78-82df-68f14b9e76b5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 21 09:45:04 crc kubenswrapper[4696]: I0321 09:45:04.007073 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09513529-8f13-4b78-82df-68f14b9e76b5-kube-api-access-px2z9" (OuterVolumeSpecName: "kube-api-access-px2z9") pod "09513529-8f13-4b78-82df-68f14b9e76b5" (UID: "09513529-8f13-4b78-82df-68f14b9e76b5"). InnerVolumeSpecName "kube-api-access-px2z9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 21 09:45:04 crc kubenswrapper[4696]: I0321 09:45:04.007555 4696 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09513529-8f13-4b78-82df-68f14b9e76b5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "09513529-8f13-4b78-82df-68f14b9e76b5" (UID: "09513529-8f13-4b78-82df-68f14b9e76b5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 21 09:45:04 crc kubenswrapper[4696]: I0321 09:45:04.103288 4696 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09513529-8f13-4b78-82df-68f14b9e76b5-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 21 09:45:04 crc kubenswrapper[4696]: I0321 09:45:04.103331 4696 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px2z9\" (UniqueName: \"kubernetes.io/projected/09513529-8f13-4b78-82df-68f14b9e76b5-kube-api-access-px2z9\") on node \"crc\" DevicePath \"\"" Mar 21 09:45:04 crc kubenswrapper[4696]: I0321 09:45:04.103343 4696 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09513529-8f13-4b78-82df-68f14b9e76b5-config-volume\") on node \"crc\" DevicePath \"\"" Mar 21 09:45:04 crc kubenswrapper[4696]: I0321 09:45:04.377956 4696 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29568105-ktz8q" event={"ID":"09513529-8f13-4b78-82df-68f14b9e76b5","Type":"ContainerDied","Data":"6ab49f60a1f4202de2276703548b4bf3dc1679e1b52e6c338a860bf8eec5bc16"} Mar 21 09:45:04 crc kubenswrapper[4696]: I0321 09:45:04.377991 4696 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ab49f60a1f4202de2276703548b4bf3dc1679e1b52e6c338a860bf8eec5bc16" Mar 21 09:45:04 crc kubenswrapper[4696]: I0321 09:45:04.378086 4696 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29568105-ktz8q" Mar 21 09:45:04 crc kubenswrapper[4696]: I0321 09:45:04.435501 4696 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568060-bj9q2"] Mar 21 09:45:04 crc kubenswrapper[4696]: I0321 09:45:04.447432 4696 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29568060-bj9q2"] Mar 21 09:45:04 crc kubenswrapper[4696]: I0321 09:45:04.517534 4696 scope.go:117] "RemoveContainer" containerID="a8834f1ba6ed59616dff8764718930d83b5348fcf1337e28a638c99401e5cb65" Mar 21 09:45:04 crc kubenswrapper[4696]: I0321 09:45:04.535982 4696 scope.go:117] "RemoveContainer" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" Mar 21 09:45:04 crc kubenswrapper[4696]: E0321 09:45:04.536405 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" Mar 21 09:45:04 crc kubenswrapper[4696]: I0321 09:45:04.547384 4696 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abbe55ca-24b3-46de-a978-aa00abe0d9e9" path="/var/lib/kubelet/pods/abbe55ca-24b3-46de-a978-aa00abe0d9e9/volumes" Mar 21 09:45:04 crc kubenswrapper[4696]: I0321 09:45:04.556513 4696 scope.go:117] "RemoveContainer" containerID="851bdfdd749137408d17275f09bb435667b969c024449803c98cfaf952273c1d" Mar 21 09:45:18 crc kubenswrapper[4696]: I0321 09:45:18.534858 4696 scope.go:117] "RemoveContainer" containerID="30fcb4559fb3a52aa03be12ff50638b70bdc474350c481c894d655a0e32665c5" Mar 21 09:45:18 crc kubenswrapper[4696]: E0321 09:45:18.536639 4696 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-z7srw_openshift-machine-config-operator(daaf227a-2305-495e-8495-a280abcd8e10)\"" pod="openshift-machine-config-operator/machine-config-daemon-z7srw" podUID="daaf227a-2305-495e-8495-a280abcd8e10" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515157464277024467 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015157464300017367 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015157452621016515 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015157452621015465 5ustar corecore